A Deep Investigation of RNN and Self-attention for the Cyrillic-Traditional Mongolian Bidirectional Conversion

Muhan Na, Rui Liu, Feilong, Guanglai Gao
{"title":"A Deep Investigation of RNN and Self-attention for the Cyrillic-Traditional Mongolian Bidirectional Conversion","authors":"Muhan Na, Rui Liu, Feilong, Guanglai Gao","doi":"10.48550/arXiv.2209.11963","DOIUrl":null,"url":null,"abstract":"Cyrillic and Traditional Mongolian are the two main members of the Mongolian writing system. The Cyrillic-Traditional Mongolian Bidirectional Conversion (CTMBC) task includes two conversion processes, including Cyrillic Mongolian to Traditional Mongolian (C2T) and Traditional Mongolian to Cyrillic Mongolian conversions (T2C). Previous researchers adopted the traditional joint sequence model, since the CTMBC task is a natural Sequence-to-Sequence (Seq2Seq) modeling problem. Recent studies have shown that Recurrent Neural Network (RNN) and Self-attention (or Transformer) based encoder-decoder models have shown significant improvement in machine translation tasks between some major languages, such as Mandarin, English, French, etc. However, an open problem remains as to whether the CTMBC quality can be improved by utilizing the RNN and Transformer models. To answer this question, this paper investigates the utility of these two powerful techniques for CTMBC task combined with agglutinative characteristics of Mongolian language. We build the encoder-decoder based CTMBC model based on RNN and Transformer respectively and compare the different network configurations deeply. The experimental results show that both RNN and Transformer models outperform the traditional joint sequence model, where the Transformer achieves the best performance. Compared with the joint sequence baseline, the word error rate (WER) of the Transformer for C2T and T2C decreased by 5.72\\% and 5.06\\% respectively.","PeriodicalId":281152,"journal":{"name":"International Conference on Neural Information Processing","volume":"19 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Neural Information Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2209.11963","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Cyrillic and Traditional Mongolian are the two main members of the Mongolian writing system. The Cyrillic-Traditional Mongolian Bidirectional Conversion (CTMBC) task includes two conversion processes, including Cyrillic Mongolian to Traditional Mongolian (C2T) and Traditional Mongolian to Cyrillic Mongolian conversions (T2C). Previous researchers adopted the traditional joint sequence model, since the CTMBC task is a natural Sequence-to-Sequence (Seq2Seq) modeling problem. Recent studies have shown that Recurrent Neural Network (RNN) and Self-attention (or Transformer) based encoder-decoder models have shown significant improvement in machine translation tasks between some major languages, such as Mandarin, English, French, etc. However, an open problem remains as to whether the CTMBC quality can be improved by utilizing the RNN and Transformer models. To answer this question, this paper investigates the utility of these two powerful techniques for CTMBC task combined with agglutinative characteristics of Mongolian language. We build the encoder-decoder based CTMBC model based on RNN and Transformer respectively and compare the different network configurations deeply. The experimental results show that both RNN and Transformer models outperform the traditional joint sequence model, where the Transformer achieves the best performance. Compared with the joint sequence baseline, the word error rate (WER) of the Transformer for C2T and T2C decreased by 5.72\% and 5.06\% respectively.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
西里尔-传统蒙古语双向转换RNN与自关注的深入研究
西里尔文和传统蒙古文是蒙古文字系统的两个主要成员。西里尔-传统蒙古语双向转换(CTMBC)任务包括西里尔蒙古语到传统蒙古语(C2T)和传统蒙古语到西里尔蒙古语(T2C)两个转换过程。由于CTMBC任务是一个自然的序列到序列(sequence -to- sequence, Seq2Seq)建模问题,以往的研究者采用传统的联合序列模型。最近的研究表明,基于循环神经网络(RNN)和自注意(或变压器)的编码器-解码器模型在一些主要语言(如汉语、英语、法语等)之间的机器翻译任务中表现出显著的改善。然而,一个悬而未决的问题仍然存在,即是否可以通过利用RNN和Transformer模型来提高ctbc的质量。为了回答这个问题,本文结合蒙古语的黏着特点,探讨了这两种强大的技术在cttmbc任务中的应用。分别建立了基于RNN和Transformer的基于编码器-解码器的CTMBC模型,并对不同的网络配置进行了深入的比较。实验结果表明,RNN和Transformer模型均优于传统的联合序列模型,其中Transformer模型的性能最好。与联合序列基线相比,变压器对C2T和T2C的字错误率(WER)分别降低了5.72%和5.06%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
On Searching for Minimal Integer Representation of Undirected Graphs Mastering Complex Coordination Through Attention-Based Dynamic Graph FIT: Frequency-based Image Translation for Domain Adaptive Object Detection Emotion Detection in Unfix-length-Context Conversation Rethinking Voxelization and Classification for 3D Object Detection
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1