{"title":"Automated Speaking Assessment of Conversation Tests with Novel Graph-based Modeling on Spoken Response Coherence","authors":"Jiun-Ting Li, Bi-Cheng Yan, Tien-Hong Lo, Yi-Cheng Wang, Yung-Chang Hsu, Berlin Chen","doi":"arxiv-2409.07064","DOIUrl":null,"url":null,"abstract":"Automated speaking assessment in conversation tests (ASAC) aims to evaluate\nthe overall speaking proficiency of an L2 (second-language) speaker in a\nsetting where an interlocutor interacts with one or more candidates. Although\nprior ASAC approaches have shown promising performance on their respective\ndatasets, there is still a dearth of research specifically focused on\nincorporating the coherence of the logical flow within a conversation into the\ngrading model. To address this critical challenge, we propose a hierarchical\ngraph model that aptly incorporates both broad inter-response interactions\n(e.g., discourse relations) and nuanced semantic information (e.g., semantic\nwords and speaker intents), which is subsequently fused with contextual\ninformation for the final prediction. Extensive experimental results on the\nNICT-JLE benchmark dataset suggest that our proposed modeling approach can\nyield considerable improvements in prediction accuracy with respect to various\nassessment metrics, as compared to some strong baselines. This also sheds light\non the importance of investigating coherence-related facets of spoken responses\nin ASAC.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"31 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computation and Language","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07064","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Automated speaking assessment in conversation tests (ASAC) aims to evaluate
the overall speaking proficiency of an L2 (second-language) speaker in a
setting where an interlocutor interacts with one or more candidates. Although
prior ASAC approaches have shown promising performance on their respective
datasets, there is still a dearth of research specifically focused on
incorporating the coherence of the logical flow within a conversation into the
grading model. To address this critical challenge, we propose a hierarchical
graph model that aptly incorporates both broad inter-response interactions
(e.g., discourse relations) and nuanced semantic information (e.g., semantic
words and speaker intents), which is subsequently fused with contextual
information for the final prediction. Extensive experimental results on the
NICT-JLE benchmark dataset suggest that our proposed modeling approach can
yield considerable improvements in prediction accuracy with respect to various
assessment metrics, as compared to some strong baselines. This also sheds light
on the importance of investigating coherence-related facets of spoken responses
in ASAC.