{"title":"Transgenerators","authors":"Arip Asadulaev, Gideon Stein, A. Filchenkov","doi":"10.1145/3446132.3446417","DOIUrl":null,"url":null,"abstract":"Pre-trained Transformers(GPT) are showed great performance in natural language generation task. This model was trained in a self-supervised manner on a large amount of text data crawled from the WEB. Such a dataset has not the highest quality, many sentences are prone to errors such as typos or grammar mistakes. As a result, text generated by GPTs consists of a lot of grammar incorrect sentences. While Transformers is also showed great performance in translation tasks, we propose the conception when a model can handle a generation and a translation task at the same time. But we propose a specific type of translation, in our method Transformer is training to translate a sentence with grammar errors to the same sentences without errors. In the full case, an incorrectly generated sentence can be corrected by the extended version of the same model, we call this type of model Transgenerator. We applied several experiments to estimate a generative power of Transgenerator based on GPT-2 architecture and the proposed method outperformed original GPT-2 model on the range of tasks","PeriodicalId":125388,"journal":{"name":"Proceedings of the 2020 3rd International Conference on Algorithms, Computing and Artificial Intelligence","volume":"18 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2020 3rd International Conference on Algorithms, Computing and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3446132.3446417","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Pre-trained Transformers(GPT) are showed great performance in natural language generation task. This model was trained in a self-supervised manner on a large amount of text data crawled from the WEB. Such a dataset has not the highest quality, many sentences are prone to errors such as typos or grammar mistakes. As a result, text generated by GPTs consists of a lot of grammar incorrect sentences. While Transformers is also showed great performance in translation tasks, we propose the conception when a model can handle a generation and a translation task at the same time. But we propose a specific type of translation, in our method Transformer is training to translate a sentence with grammar errors to the same sentences without errors. In the full case, an incorrectly generated sentence can be corrected by the extended version of the same model, we call this type of model Transgenerator. We applied several experiments to estimate a generative power of Transgenerator based on GPT-2 architecture and the proposed method outperformed original GPT-2 model on the range of tasks