{"title":"Improving Abstractive Summarization by Transfer Learning with Adaptive Document Selection","authors":"Masato Shirai, Kei Wakabayashi","doi":"10.52731/ijscai.v7.i2.701","DOIUrl":null,"url":null,"abstract":"ive document summarization based on neural networks is a promising approach to generate a flexible summary but requires a large amount of training data.While transfer learning can address this issue, there is a potential concern about the negative transfer effect that deteriorates the performance when we use training documents irrelevant to the target domain, which has not been explicitly explored in document summarization tasks.In this paper, we propose a method that selects training documents from the source domain that are expected to be useful for the target summarization.The proposed method is based on the similarity of word distributions between each source document and a set of target documents.We further propose an adaptive approach that builds a custom-made summarization model for each test document by selecting source documents similar to the test document.In the experiment, we confirmed that the negative transfer actually happens also in the document summarization tasks.Additionally, we show that the proposed method effectively avoids the negative transfer issue and improves summarization performance.","PeriodicalId":495454,"journal":{"name":"International journal of smart computing and artificial intelligence","volume":"2013 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International journal of smart computing and artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.52731/ijscai.v7.i2.701","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
ive document summarization based on neural networks is a promising approach to generate a flexible summary but requires a large amount of training data.While transfer learning can address this issue, there is a potential concern about the negative transfer effect that deteriorates the performance when we use training documents irrelevant to the target domain, which has not been explicitly explored in document summarization tasks.In this paper, we propose a method that selects training documents from the source domain that are expected to be useful for the target summarization.The proposed method is based on the similarity of word distributions between each source document and a set of target documents.We further propose an adaptive approach that builds a custom-made summarization model for each test document by selecting source documents similar to the test document.In the experiment, we confirmed that the negative transfer actually happens also in the document summarization tasks.Additionally, we show that the proposed method effectively avoids the negative transfer issue and improves summarization performance.