Learning to Correct Erroneous Words for Document Grounded Conversations

Junyan Qiu, Haidong Zhang, Yiping Yang
{"title":"Learning to Correct Erroneous Words for Document Grounded Conversations","authors":"Junyan Qiu, Haidong Zhang, Yiping Yang","doi":"10.1145/3587828.3587883","DOIUrl":null,"url":null,"abstract":"Document grounded conversation (DGC) aims to generate informative responses when talking about a document. It is normally formulated as a sequence-to-sequence (Seq2seq) learning problem, which directly maps source sequences, i.e., the context and background documents, to the target sequence, i.e., the response. These responses are normally used as the final output without further polishing, which may suffer from the global information loss owing to the auto-regression paradigm. To tackle this problem, some researches designed two-pass generation to improve the quality of responses. However, these approaches lack the capability of distinguishing inappropriate words in the first pass, which may maintain the erroneous words while rewrite the correct ones. In this paper, we design a scheduled error correction network (SECN) with multiple generation passes to explicitly locate and rewrite the erroneous words in previous passes. Specifically, a discriminator is employed to distinguish erroneous words which are further revised by a refiner. Moreover, we also apply curriculum learning with reasonable learning schedule to train our model from easy to hard conversations, where the complexity is measured by the number of decoding passes. We conduct comprehensive experiments on a public document grounded conversation dataset, Wizard-of-Wikipedia, and the results demonstrate significant promotions over several strong benchmarks.","PeriodicalId":340917,"journal":{"name":"Proceedings of the 2023 12th International Conference on Software and Computer Applications","volume":"221 1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2023 12th International Conference on Software and Computer Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3587828.3587883","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Document grounded conversation (DGC) aims to generate informative responses when talking about a document. It is normally formulated as a sequence-to-sequence (Seq2seq) learning problem, which directly maps source sequences, i.e., the context and background documents, to the target sequence, i.e., the response. These responses are normally used as the final output without further polishing, which may suffer from the global information loss owing to the auto-regression paradigm. To tackle this problem, some researches designed two-pass generation to improve the quality of responses. However, these approaches lack the capability of distinguishing inappropriate words in the first pass, which may maintain the erroneous words while rewrite the correct ones. In this paper, we design a scheduled error correction network (SECN) with multiple generation passes to explicitly locate and rewrite the erroneous words in previous passes. Specifically, a discriminator is employed to distinguish erroneous words which are further revised by a refiner. Moreover, we also apply curriculum learning with reasonable learning schedule to train our model from easy to hard conversations, where the complexity is measured by the number of decoding passes. We conduct comprehensive experiments on a public document grounded conversation dataset, Wizard-of-Wikipedia, and the results demonstrate significant promotions over several strong benchmarks.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
学习纠正基于文件的对话中的错误单词
基于文档的对话(DGC)的目的是在谈论文档时生成信息丰富的响应。它通常被表述为序列到序列(Seq2seq)学习问题,它直接将源序列(即上下文和背景文档)映射到目标序列(即响应)。这些响应通常用作最终输出,而不进行进一步的修饰,这可能会由于自回归范式而遭受全局信息损失。为了解决这个问题,一些研究设计了两遍生成来提高响应的质量。然而,这些方法缺乏第一次识别不恰当词的能力,这可能会保留错误的词,而重写正确的词。在本文中,我们设计了一个具有多个生成通道的定时纠错网络(SECN),以显式定位和重写先前通道中的错误单词。具体地说,使用一个鉴别器来区分错别字,这些错别字由精炼器进一步修正。此外,我们还应用课程学习和合理的学习计划来训练我们的模型,从简单到困难的对话,其中复杂性是通过解码通过的次数来衡量的。我们在基于对话数据集Wizard-of-Wikipedia的公共文档上进行了全面的实验,结果表明在几个强大的基准测试中有显著的提升。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
The Use of Dynamic n-Gram to Enhance TF-IDF Features Extraction for Bahasa Indonesia Cyberbullying Classification Development of IT Equipment Management Methodology based on Carbon Emission and End-of-Life Period with A Design Thinking Approach: Case Study: Bandung Institute of Technology Formal Specification and Model Checking of Raft Leader Election in Maude* An Ontology-based Modeling for Classifying Risk of Suicidal Behavior String Figure Simulation with Multiresolution Wire Model
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1