{"title":"Examining Manual and Automatic MT Evaluation: A Grey Relational Analysis for Chinese-Portuguese Translation Quality","authors":"Yuqi Sun, Lap-Man Hoi, S. Im","doi":"10.1109/CCAI57533.2023.10201322","DOIUrl":null,"url":null,"abstract":"This study investigates the relationship between manual and automatic machine translation evaluation methodologies by employing Grey Correlation Analysis (GRA) to assess the correlation between Chinese-Portuguese machine translation outputs’ BLEU scores and human evaluation scores based on a proposed evaluation index system. The research aims to provide insights into the factors impacting machine translation quality and the most relevant linguistic dimensions in translation evaluation. The findings reveal that “usability” and “adequacy” exhibit the highest correlation with BLEU scores, while “Semantics” ranks highest among the ten manual evaluation indicators, when correlated with the aggregated human evaluation results, followed by “Correction” (correct information) and “Omission and/or Addition”. The findings contribute to the field of machine translation evaluation by illuminating the complex relationship between manual and automatic evaluation techniques and guiding future improvements in machine translation systems and evaluation methodologies.","PeriodicalId":285760,"journal":{"name":"2023 IEEE 3rd International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"86 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 3rd International Conference on Computer Communication and Artificial Intelligence (CCAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCAI57533.2023.10201322","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This study investigates the relationship between manual and automatic machine translation evaluation methodologies by employing Grey Correlation Analysis (GRA) to assess the correlation between Chinese-Portuguese machine translation outputs’ BLEU scores and human evaluation scores based on a proposed evaluation index system. The research aims to provide insights into the factors impacting machine translation quality and the most relevant linguistic dimensions in translation evaluation. The findings reveal that “usability” and “adequacy” exhibit the highest correlation with BLEU scores, while “Semantics” ranks highest among the ten manual evaluation indicators, when correlated with the aggregated human evaluation results, followed by “Correction” (correct information) and “Omission and/or Addition”. The findings contribute to the field of machine translation evaluation by illuminating the complex relationship between manual and automatic evaluation techniques and guiding future improvements in machine translation systems and evaluation methodologies.