{"title":"GLSE: Global-Local Selective Encoding for Response Generation in Neural Conversation Model","authors":"Hongli Wang, Jiangtao Ren","doi":"10.1109/ICTAI.2019.00166","DOIUrl":null,"url":null,"abstract":"How to generate relevant and informative response is one of the core topics in response generation area. Following the task formulation of neural machine translation, previous works mainly consider response generation task as a mapping from a source sentence to a target sentence. However, the dialogue model tends to generate safe, commonplace responses (e.g., I don't know) regardless of the input, when learning to maximize the likelihood of response for the given message in an almost loss-less manner just like MT. Different from existing works, we propose a Global-Local Selective Encoding model (GLSE) to extend the seq2seq framework to generate more relevant and informative responses. Specifically, two types of selective gate network are introduced in this work: (i) A local selective word-sentence gate is added after encoding phase of Seq2Seq learning framework, which learns to tailor the original message information and generates a selected input representation. (ii) A global selective bidirectional-context gate is set to control the bidirectional information flow from a BiGRU based encoder to decoder. Empirical studies indicate the advantage of our model over several classical and strong baselines.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"87 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTAI.2019.00166","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
How to generate relevant and informative response is one of the core topics in response generation area. Following the task formulation of neural machine translation, previous works mainly consider response generation task as a mapping from a source sentence to a target sentence. However, the dialogue model tends to generate safe, commonplace responses (e.g., I don't know) regardless of the input, when learning to maximize the likelihood of response for the given message in an almost loss-less manner just like MT. Different from existing works, we propose a Global-Local Selective Encoding model (GLSE) to extend the seq2seq framework to generate more relevant and informative responses. Specifically, two types of selective gate network are introduced in this work: (i) A local selective word-sentence gate is added after encoding phase of Seq2Seq learning framework, which learns to tailor the original message information and generates a selected input representation. (ii) A global selective bidirectional-context gate is set to control the bidirectional information flow from a BiGRU based encoder to decoder. Empirical studies indicate the advantage of our model over several classical and strong baselines.