Non-Autoregressive Translation Algorithm Based on LLM Knowledge Distillation in English Corpus

IF 2 Q3 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Engineering reports : open access Pub Date : 2024-12-08 DOI:10.1002/eng2.13077
Fang Ju, Weihui Wang
{"title":"Non-Autoregressive Translation Algorithm Based on LLM Knowledge Distillation in English Corpus","authors":"Fang Ju,&nbsp;Weihui Wang","doi":"10.1002/eng2.13077","DOIUrl":null,"url":null,"abstract":"<p>Although significant advancements have been made in the quality of machine translation by large-scale language models, their high computational costs and resource consumption have hindered their widespread adoption in practical applications. So this research introduces an English corpus-based machine translation algorithm that leverages knowledge distillation from large language model, with the goal of enhancing translation quality and reducing the computational demands of the model. Initially, we conducted a thorough analysis of the English corpus to identify prevalent language patterns and structures. Following this, we developed a knowledge distillation approach that transfers the translation expertise of a large teacher model to a smaller student model, thereby achieving increased translation accuracy and efficiency. We designed a dynamic temperature hyperparameter distillation strategy that effectively enhances the precision of translations. In the experimental phase, we utilized several standard English corpora to train and assess our algorithm. The findings indicate that, compared to current machine translation systems, our method significantly reduces the need for computational resources while preserving translation quality.</p>","PeriodicalId":72922,"journal":{"name":"Engineering reports : open access","volume":"7 1","pages":""},"PeriodicalIF":2.0000,"publicationDate":"2024-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/eng2.13077","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering reports : open access","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/eng2.13077","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0

Abstract

Although significant advancements have been made in the quality of machine translation by large-scale language models, their high computational costs and resource consumption have hindered their widespread adoption in practical applications. So this research introduces an English corpus-based machine translation algorithm that leverages knowledge distillation from large language model, with the goal of enhancing translation quality and reducing the computational demands of the model. Initially, we conducted a thorough analysis of the English corpus to identify prevalent language patterns and structures. Following this, we developed a knowledge distillation approach that transfers the translation expertise of a large teacher model to a smaller student model, thereby achieving increased translation accuracy and efficiency. We designed a dynamic temperature hyperparameter distillation strategy that effectively enhances the precision of translations. In the experimental phase, we utilized several standard English corpora to train and assess our algorithm. The findings indicate that, compared to current machine translation systems, our method significantly reduces the need for computational resources while preserving translation quality.

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于LLM知识蒸馏的英语语料非自回归翻译算法
尽管大规模语言模型在机器翻译质量方面取得了重大进展,但其高昂的计算成本和资源消耗阻碍了其在实际应用中的广泛采用。因此,本研究引入了一种基于英语语料库的机器翻译算法,该算法利用大型语言模型的知识蒸馏,以提高翻译质量并减少模型的计算需求。首先,我们对英语语料库进行了全面的分析,以确定流行的语言模式和结构。在此之后,我们开发了一种知识蒸馏方法,将大型教师模型的翻译专业知识转移到较小的学生模型,从而提高了翻译的准确性和效率。设计了一种动态温度超参数蒸馏策略,有效地提高了翻译精度。在实验阶段,我们使用了几个标准的英语语料库来训练和评估我们的算法。研究结果表明,与现有的机器翻译系统相比,我们的方法在保持翻译质量的同时显著减少了对计算资源的需求。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
5.10
自引率
0.00%
发文量
0
审稿时长
19 weeks
期刊最新文献
Advanced Statistical Characterization and Correlation Analysis of Process Performance Indicators for Optimized Engineering Decisions RogueGPT: Unleashing Jailbreak Prompts on LLMs A Review on Explainable, Federated Multimodal AI for Heart Disease Detection Using ECG, Cardiac Imaging, and Electronic Health Records Numerical and Intelligent Modeling of MHD Casson Nanofluid Heat Transfer in Fractal Porous Cavities for Energy Systems A Smartphone and Web-Based Automated Platform for Segmenting Urinary Tract Infection Using a Deep Learning-Based Approach
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1