分层编码器-解码器总结模型与讲师长学术论文

Jianling Li, Wuhang Lin, Shasha Li, Jie Yu, Jun Ma
{"title":"分层编码器-解码器总结模型与讲师长学术论文","authors":"Jianling Li, Wuhang Lin, Shasha Li, Jie Yu, Jun Ma","doi":"10.1109/ICCCS52626.2021.9449262","DOIUrl":null,"url":null,"abstract":"Summary models, whether extractive or abstractive, have achieved great success recently. For long academic papers, the abstractive model with the encoder-decoder architecture mainly only relies on the attentional context vector for generation, unlike humans who have already mastered the salient information of the source text to have full control over what to write. While the extracted sentences always contain the correct and salient information which can be used to control the abstraction process. Therefore, based on a hierarchical encoder-decoder architecture specifically for academic papers, we proposed a summary model with an Instructor, an encoder in essence by taking the guiding sentences as the input to further control the generating process. In the encoder part, the final hidden state from Instructor is directly added to the basic hierarchical hidden state from the encoder. Experimental results on arXiv/PubMed show that the only encoder-improved model can generate better abstract. In the decoder part, the context vector from Instructor is integrated with the original discourse-aware context vector for the generation. The results show that Instructor is effective for control and our model can generate a more accurate and fluent abstract with significantly higher ROUGE values.","PeriodicalId":376290,"journal":{"name":"2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Hierarchical Encoder-Decoder Summary Model with an Instructor for Long Academic Papers\",\"authors\":\"Jianling Li, Wuhang Lin, Shasha Li, Jie Yu, Jun Ma\",\"doi\":\"10.1109/ICCCS52626.2021.9449262\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Summary models, whether extractive or abstractive, have achieved great success recently. For long academic papers, the abstractive model with the encoder-decoder architecture mainly only relies on the attentional context vector for generation, unlike humans who have already mastered the salient information of the source text to have full control over what to write. While the extracted sentences always contain the correct and salient information which can be used to control the abstraction process. Therefore, based on a hierarchical encoder-decoder architecture specifically for academic papers, we proposed a summary model with an Instructor, an encoder in essence by taking the guiding sentences as the input to further control the generating process. In the encoder part, the final hidden state from Instructor is directly added to the basic hierarchical hidden state from the encoder. Experimental results on arXiv/PubMed show that the only encoder-improved model can generate better abstract. In the decoder part, the context vector from Instructor is integrated with the original discourse-aware context vector for the generation. The results show that Instructor is effective for control and our model can generate a more accurate and fluent abstract with significantly higher ROUGE values.\",\"PeriodicalId\":376290,\"journal\":{\"name\":\"2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS)\",\"volume\":\"25 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-04-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCCS52626.2021.9449262\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCCS52626.2021.9449262","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

摘要模型,无论是抽取的还是抽象的,最近都取得了巨大的成功。对于长篇学术论文,具有编码器-解码器架构的抽象模型主要只依赖于注意力上下文向量进行生成,不像人类已经掌握了源文本的显著信息,完全可以控制写什么。而提取出来的句子总是包含着正确的、显著的信息,这些信息可以用来控制提取过程。因此,基于专门针对学术论文的分层编码器-解码器架构,我们提出了一个带有讲师的摘要模型,本质上是一个编码器,将引导句作为输入,进一步控制生成过程。在编码器部分,最终隐藏状态直接添加到编码器的基本层次隐藏状态。在arXiv/PubMed上的实验结果表明,只有编码改进的模型才能生成更好的摘要。在解码器部分,将来自讲师的上下文向量与原始话语感知上下文向量相结合进行生成。结果表明,该方法具有较好的控制效果,模型可以生成更准确、更流畅的抽象,且ROUGE值显著提高。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Hierarchical Encoder-Decoder Summary Model with an Instructor for Long Academic Papers
Summary models, whether extractive or abstractive, have achieved great success recently. For long academic papers, the abstractive model with the encoder-decoder architecture mainly only relies on the attentional context vector for generation, unlike humans who have already mastered the salient information of the source text to have full control over what to write. While the extracted sentences always contain the correct and salient information which can be used to control the abstraction process. Therefore, based on a hierarchical encoder-decoder architecture specifically for academic papers, we proposed a summary model with an Instructor, an encoder in essence by taking the guiding sentences as the input to further control the generating process. In the encoder part, the final hidden state from Instructor is directly added to the basic hierarchical hidden state from the encoder. Experimental results on arXiv/PubMed show that the only encoder-improved model can generate better abstract. In the decoder part, the context vector from Instructor is integrated with the original discourse-aware context vector for the generation. The results show that Instructor is effective for control and our model can generate a more accurate and fluent abstract with significantly higher ROUGE values.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
A Method of Measuring Data Fusion Based on EMBET Real Time Noise Power Estimation for Single Carrier Frequency Domain Equalization The CPDA Detector for the MIMO OCDM System A Cooperative Search Algorithm Based on Improved Particle Swarm Optimization Decision for UAV Swarm A Network Topology Awareness Based Probabilistic Broadcast Protocol for Data Transmission in Mobile Ad Hoc Networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1