An Extensive Study on Pretrained Models for Natural Language Processing Based on Transformers

M. Ramprasath, K. Dhanasekaran, T. Karthick, R. Velumani, P. Sudhakaran
{"title":"An Extensive Study on Pretrained Models for Natural Language Processing Based on Transformers","authors":"M. Ramprasath, K. Dhanasekaran, T. Karthick, R. Velumani, P. Sudhakaran","doi":"10.1109/ICEARS53579.2022.9752241","DOIUrl":null,"url":null,"abstract":"In recent years, Pretraining Language Models based on Transformers -Natural Language Processing (PLMT-NLP) have been highly successful in nearly every NLP task. In the beginning, Generative Pre-trained model-based Transformer, BERT- Bidirectional Encoder model Representations using Transformers was used to develop these models. Models constructed on transformers, Self-supervise knowledge acquiring, and transfer learning establish the foundation of these designs. Transformed-based pre-trained models acquire common linguistic illustrations from vast amounts of textual information through self-supervised model and apply this information to downstream tasks. To eliminate the need to retrain downstream models, these models provide a solid foundation of knowledge. In this paper, the enhanced learning on PLMT-NLP has been discussed. Initially, a quick introduction to self-supervised learning is presented, then diverse core concepts used in PLMT-NLP are explained. Furthermore, a list of relevant libraries for working with PLMT-NLP has been provided. Lastly, the paper discusses about the upcoming research directions that will further improve these models. Because of its thoroughness and relevance to current PLMT-NLP developments, this survey study will positively serve as a valuable resource for those seeking to understand both basic ideas and new developments better.","PeriodicalId":252961,"journal":{"name":"2022 International Conference on Electronics and Renewable Systems (ICEARS)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Electronics and Renewable Systems (ICEARS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICEARS53579.2022.9752241","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

In recent years, Pretraining Language Models based on Transformers -Natural Language Processing (PLMT-NLP) have been highly successful in nearly every NLP task. In the beginning, Generative Pre-trained model-based Transformer, BERT- Bidirectional Encoder model Representations using Transformers was used to develop these models. Models constructed on transformers, Self-supervise knowledge acquiring, and transfer learning establish the foundation of these designs. Transformed-based pre-trained models acquire common linguistic illustrations from vast amounts of textual information through self-supervised model and apply this information to downstream tasks. To eliminate the need to retrain downstream models, these models provide a solid foundation of knowledge. In this paper, the enhanced learning on PLMT-NLP has been discussed. Initially, a quick introduction to self-supervised learning is presented, then diverse core concepts used in PLMT-NLP are explained. Furthermore, a list of relevant libraries for working with PLMT-NLP has been provided. Lastly, the paper discusses about the upcoming research directions that will further improve these models. Because of its thoroughness and relevance to current PLMT-NLP developments, this survey study will positively serve as a valuable resource for those seeking to understand both basic ideas and new developments better.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于变形器的自然语言处理预训练模型的广泛研究
近年来,基于变形器-自然语言处理(PLMT-NLP)的预训练语言模型在几乎所有的NLP任务中都取得了巨大的成功。首先,基于生成预训练模型的Transformer, BERT-使用Transformer的双向编码器模型表示用于开发这些模型。基于变压器的模型构建、自我监督知识获取和迁移学习为这些设计奠定了基础。基于转换的预训练模型通过自监督模型从大量文本信息中获取共同的语言插图,并将这些信息应用于下游任务。为了消除对下游模型进行再培训的需要,这些模型提供了坚实的知识基础。本文对PLMT-NLP的强化学习进行了讨论。首先,介绍了自我监督学习,然后解释了PLMT-NLP中使用的各种核心概念。此外,还提供了使用PLMT-NLP的相关库的列表。最后,对未来的研究方向进行了展望,以进一步完善这些模型。由于其彻底性和与当前PLMT-NLP发展的相关性,这项调查研究将积极地为那些寻求更好地理解基本思想和新发展的人提供宝贵的资源。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
The Solar Tracker Using Micro-controller "Core Strength" of Dance Lala Training Considering the Body Motion Tracking Video and Predictive Model Textile Antenna –Structure, Material and Applications Automated Classification of Atherosclerosis in Coronary Computed Tomography Angiography Images Based on Radiomics Study Using Automatic Machine Learning Cryptocurrency Exchange Rate Prediction using ARIMA Model on Real Time Data
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1