{"title":"基于对齐偏移量的同步机器翻译自适应训练","authors":"Qiqi Liang, Yanjun Liu, Fandong Meng, Jinan Xu, Yufeng Chen, Jie Zhou","doi":"10.1109/ICNLP58431.2023.00035","DOIUrl":null,"url":null,"abstract":"Given incomplete source sentences as inputs, it is generally difficult for Simultaneous Machine Translation (SiMT) models to generate a target token once its aligned source tokens are absent. How to measure such difficulty and further conduct adaptive training for SiMT models are not sufficiently studied. In this paper, we propose a new metric named alignment offset (AO) to quantify the learning difficulty of target tokens for SiMT models. Given a target token, its AO is calculated by the offset between its aligned source tokens and the already received source tokens. Furthermore, we design two AO-based adaptive training methods to improve the training of SiMT models. Firstly, we introduce token-level curriculum learning based on AO, which progressively switches the training process from easy target tokens to difficult ones. Secondly, we assign an appropriate weight to the training loss of each target token according to its AO. Experimental results on four datasets demonstrate that our methods significantly and consistently outperform all the strong baselines.","PeriodicalId":53637,"journal":{"name":"Icon","volume":"28 1","pages":"157-164"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Alignment Offset Based Adaptive Training for Simultaneous Machine Translation\",\"authors\":\"Qiqi Liang, Yanjun Liu, Fandong Meng, Jinan Xu, Yufeng Chen, Jie Zhou\",\"doi\":\"10.1109/ICNLP58431.2023.00035\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Given incomplete source sentences as inputs, it is generally difficult for Simultaneous Machine Translation (SiMT) models to generate a target token once its aligned source tokens are absent. How to measure such difficulty and further conduct adaptive training for SiMT models are not sufficiently studied. In this paper, we propose a new metric named alignment offset (AO) to quantify the learning difficulty of target tokens for SiMT models. Given a target token, its AO is calculated by the offset between its aligned source tokens and the already received source tokens. Furthermore, we design two AO-based adaptive training methods to improve the training of SiMT models. Firstly, we introduce token-level curriculum learning based on AO, which progressively switches the training process from easy target tokens to difficult ones. Secondly, we assign an appropriate weight to the training loss of each target token according to its AO. Experimental results on four datasets demonstrate that our methods significantly and consistently outperform all the strong baselines.\",\"PeriodicalId\":53637,\"journal\":{\"name\":\"Icon\",\"volume\":\"28 1\",\"pages\":\"157-164\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Icon\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICNLP58431.2023.00035\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"Arts and Humanities\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Icon","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNLP58431.2023.00035","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Arts and Humanities","Score":null,"Total":0}
Alignment Offset Based Adaptive Training for Simultaneous Machine Translation
Given incomplete source sentences as inputs, it is generally difficult for Simultaneous Machine Translation (SiMT) models to generate a target token once its aligned source tokens are absent. How to measure such difficulty and further conduct adaptive training for SiMT models are not sufficiently studied. In this paper, we propose a new metric named alignment offset (AO) to quantify the learning difficulty of target tokens for SiMT models. Given a target token, its AO is calculated by the offset between its aligned source tokens and the already received source tokens. Furthermore, we design two AO-based adaptive training methods to improve the training of SiMT models. Firstly, we introduce token-level curriculum learning based on AO, which progressively switches the training process from easy target tokens to difficult ones. Secondly, we assign an appropriate weight to the training loss of each target token according to its AO. Experimental results on four datasets demonstrate that our methods significantly and consistently outperform all the strong baselines.