{"title":"Multi-turn Dialogue Model Based on the Improved Hierarchical Recurrent Attention Network","authors":"Jiawei Miao, Jiansheng Wu","doi":"10.31534/engmod.2021.2.ri.02d","DOIUrl":null,"url":null,"abstract":"When considering the multi-turn dialogue systems, the model needs to generate a natural and contextual response. At present, HRAN, one of the most advanced models for multi-turn dialogue problems, uses a hierarchical recurrent encoder-decoder combined with a hierarchical attention mechanism. However, for complex conversations, the traditional attention-based RNN does not fully understand the context, which results in attention to the wrong context that generates irrelevant responses. To solve this problem, we proposed an improved hierarchical recurrent attention network, a self-attention network (HSAN), instead of RNN, to learn word representations and utterances representations. Empirical studies on both Chinese and English datasets show that the proposed model has achieved significant improvement.","PeriodicalId":35748,"journal":{"name":"International Journal for Engineering Modelling","volume":"63 4","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2021-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal for Engineering Modelling","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.31534/engmod.2021.2.ri.02d","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Engineering","Score":null,"Total":0}
引用次数: 1
Abstract
When considering the multi-turn dialogue systems, the model needs to generate a natural and contextual response. At present, HRAN, one of the most advanced models for multi-turn dialogue problems, uses a hierarchical recurrent encoder-decoder combined with a hierarchical attention mechanism. However, for complex conversations, the traditional attention-based RNN does not fully understand the context, which results in attention to the wrong context that generates irrelevant responses. To solve this problem, we proposed an improved hierarchical recurrent attention network, a self-attention network (HSAN), instead of RNN, to learn word representations and utterances representations. Empirical studies on both Chinese and English datasets show that the proposed model has achieved significant improvement.
期刊介绍:
Engineering Modelling is a refereed international journal providing an up-to-date reference for the engineers and researchers engaged in computer aided analysis, design and research in the fields of computational mechanics, numerical methods, software develop-ment and engineering modelling.