{"title":"基于变压器双向编码器表示的文本摘要生成研究","authors":"Wen Kai, Zhou Lingyu","doi":"10.1109/ITCA52113.2020.00074","DOIUrl":null,"url":null,"abstract":"For Chinese automatic summarization, most of the generation methods are extractive, and the generative summary is not smooth, incoherent, and covers incomplete information. Compared with the traditional sequence-to-sequence model, Generative Adversarial Network (GAN) uses a reinforcement learning strategy The use of discriminator to guide generation has achieved good results in text generation. This paper proposes a pre-training method based on Bidirectional Encoder Representation from Transformers (BERT) and combined with LeakGAN model to generate abstracts. Firstly, using the bidirectional encoding characteristics of the BERT model, it can retain the original information well, and has a better effect when extracting features of words in the context to generate high-quality word vectors; secondly, for the current supervised generative model Both have the training problem of maximum likelihood estimation. This article uses the LeakGAN model that can decompose the task into different levels of sub-strategies, and uses hierarchical reinforcement learning to solve the characteristics of sparse rewards and generate a more accurate summary.","PeriodicalId":103309,"journal":{"name":"2020 2nd International Conference on Information Technology and Computer Application (ITCA)","volume":"108 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Research on Text Summary Generation Based on Bidirectional Encoder Representation from Transformers\",\"authors\":\"Wen Kai, Zhou Lingyu\",\"doi\":\"10.1109/ITCA52113.2020.00074\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"For Chinese automatic summarization, most of the generation methods are extractive, and the generative summary is not smooth, incoherent, and covers incomplete information. Compared with the traditional sequence-to-sequence model, Generative Adversarial Network (GAN) uses a reinforcement learning strategy The use of discriminator to guide generation has achieved good results in text generation. This paper proposes a pre-training method based on Bidirectional Encoder Representation from Transformers (BERT) and combined with LeakGAN model to generate abstracts. Firstly, using the bidirectional encoding characteristics of the BERT model, it can retain the original information well, and has a better effect when extracting features of words in the context to generate high-quality word vectors; secondly, for the current supervised generative model Both have the training problem of maximum likelihood estimation. This article uses the LeakGAN model that can decompose the task into different levels of sub-strategies, and uses hierarchical reinforcement learning to solve the characteristics of sparse rewards and generate a more accurate summary.\",\"PeriodicalId\":103309,\"journal\":{\"name\":\"2020 2nd International Conference on Information Technology and Computer Application (ITCA)\",\"volume\":\"108 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 2nd International Conference on Information Technology and Computer Application (ITCA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ITCA52113.2020.00074\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 2nd International Conference on Information Technology and Computer Application (ITCA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITCA52113.2020.00074","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Research on Text Summary Generation Based on Bidirectional Encoder Representation from Transformers
For Chinese automatic summarization, most of the generation methods are extractive, and the generative summary is not smooth, incoherent, and covers incomplete information. Compared with the traditional sequence-to-sequence model, Generative Adversarial Network (GAN) uses a reinforcement learning strategy The use of discriminator to guide generation has achieved good results in text generation. This paper proposes a pre-training method based on Bidirectional Encoder Representation from Transformers (BERT) and combined with LeakGAN model to generate abstracts. Firstly, using the bidirectional encoding characteristics of the BERT model, it can retain the original information well, and has a better effect when extracting features of words in the context to generate high-quality word vectors; secondly, for the current supervised generative model Both have the training problem of maximum likelihood estimation. This article uses the LeakGAN model that can decompose the task into different levels of sub-strategies, and uses hierarchical reinforcement learning to solve the characteristics of sparse rewards and generate a more accurate summary.