首页 > 最新文献

Transactions of the Association for Computational Linguistics最新文献

英文 中文
Shared Lexical Items as Triggers of Code Switching 共享词条是代码转换的触发器
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-29 DOI: 10.1162/tacl_a_00613
S. Wintner, Safaa Shehadi, Yuli Zeira, Doreen Osmelak, Yuval Nov
Abstract Why do bilingual speakers code-switch (mix their two languages)? Among the several theories that attempt to explain this natural and ubiquitous phenomenon, the triggering hypothesis relates code-switching to the presence of lexical triggers, specifically cognates and proper names, adjacent to the switch point. We provide a fuller, more nuanced and refined exploration of the triggering hypothesis, based on five large datasets in three language pairs, reflecting both spoken and written bilingual interactions. Our results show that words that are assumed to reside in a mental lexicon shared by both languages indeed trigger code-switching, that the tendency to switch depends on the distance of the trigger from the switch point and on whether the trigger precedes or succeeds the switch, but not on the etymology of the trigger words. We thus provide strong, robust, evidence-based confirmation to several hypotheses on the relationships between lexical triggers and code-switching.
摘要 为什么说二语的人会进行语码转换(混合使用两种语言)?在试图解释这一无处不在的自然现象的几种理论中,"触发假说"(triggering hypothesis)将代码转换与词汇触发器(特别是与转换点相邻的同义词和专有名词)的存在联系起来。我们基于三个语言对的五个大型数据集,对触发假说进行了更全面、更细致、更精炼的探讨,这些数据集反映了口语和书面双语互动。我们的研究结果表明,假设存在于两种语言共享的心理词典中的词确实会触发代码转换,转换倾向取决于触发词与转换点的距离,以及触发词是在转换之前还是之后,但与触发词的词源无关。因此,我们为关于词汇触发和代码转换之间关系的几个假设提供了强有力的、基于证据的证实。
{"title":"Shared Lexical Items as Triggers of Code Switching","authors":"S. Wintner, Safaa Shehadi, Yuli Zeira, Doreen Osmelak, Yuval Nov","doi":"10.1162/tacl_a_00613","DOIUrl":"https://doi.org/10.1162/tacl_a_00613","url":null,"abstract":"Abstract Why do bilingual speakers code-switch (mix their two languages)? Among the several theories that attempt to explain this natural and ubiquitous phenomenon, the triggering hypothesis relates code-switching to the presence of lexical triggers, specifically cognates and proper names, adjacent to the switch point. We provide a fuller, more nuanced and refined exploration of the triggering hypothesis, based on five large datasets in three language pairs, reflecting both spoken and written bilingual interactions. Our results show that words that are assumed to reside in a mental lexicon shared by both languages indeed trigger code-switching, that the tendency to switch depends on the distance of the trigger from the switch point and on whether the trigger precedes or succeeds the switch, but not on the etymology of the trigger words. We thus provide strong, robust, evidence-based confirmation to several hypotheses on the relationships between lexical triggers and code-switching.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"18 1","pages":"1471-1484"},"PeriodicalIF":10.9,"publicationDate":"2023-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139348580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Can Authorship Representation Learning Capture Stylistic Features? 作者表征学习能否捕捉文体特征?
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-22 DOI: 10.1162/tacl_a_00610
Andrew Wang, Cristina Aggazzotti, R. Kotula, Rafael A. Rivera Soto, M. Bishop, Nicholas Andrews
Abstract Automatically disentangling an author’s style from the content of their writing is a longstanding and possibly insurmountable problem in computational linguistics. At the same time, the availability of large text corpora furnished with author labels has recently enabled learning authorship representations in a purely data-driven manner for authorship attribution, a task that ostensibly depends to a greater extent on encoding writing style than encoding content. However, success on this surrogate task does not ensure that such representations capture writing style since authorship could also be correlated with other latent variables, such as topic. In an effort to better understand the nature of the information these representations convey, and specifically to validate the hypothesis that they chiefly encode writing style, we systematically probe these representations through a series of targeted experiments. The results of these experiments suggest that representations learned for the surrogate authorship prediction task are indeed sensitive to writing style. As a consequence, authorship representations may be expected to be robust to certain kinds of data shift, such as topic drift over time. Additionally, our findings may open the door to downstream applications that require stylistic representations, such as style transfer.
摘要 在计算语言学中,自动将作者的风格与其写作内容区分开来是一个长期存在且可能无法解决的问题。与此同时,由于可以获得带有作者标签的大型文本语料库,因此最近能够以纯数据驱动的方式学习作者身份表征,用于作者身份归属,这项任务表面上看更依赖于对写作风格的编码,而不是对内容的编码。然而,这种代用任务的成功并不能确保此类表征能够捕捉到写作风格,因为作者身份还可能与其他潜在变量(如主题)相关。为了更好地理解这些表征所传达信息的性质,特别是验证它们主要编码写作风格的假设,我们通过一系列有针对性的实验对这些表征进行了系统的探究。这些实验的结果表明,在代理作者身份预测任务中学习到的表征确实对写作风格很敏感。因此,作者身份表征有望对某些类型的数据转移(如主题随时间的漂移)具有稳健性。此外,我们的发现可能会为需要风格表征的下游应用(如风格转移)打开大门。
{"title":"Can Authorship Representation Learning Capture Stylistic Features?","authors":"Andrew Wang, Cristina Aggazzotti, R. Kotula, Rafael A. Rivera Soto, M. Bishop, Nicholas Andrews","doi":"10.1162/tacl_a_00610","DOIUrl":"https://doi.org/10.1162/tacl_a_00610","url":null,"abstract":"Abstract Automatically disentangling an author’s style from the content of their writing is a longstanding and possibly insurmountable problem in computational linguistics. At the same time, the availability of large text corpora furnished with author labels has recently enabled learning authorship representations in a purely data-driven manner for authorship attribution, a task that ostensibly depends to a greater extent on encoding writing style than encoding content. However, success on this surrogate task does not ensure that such representations capture writing style since authorship could also be correlated with other latent variables, such as topic. In an effort to better understand the nature of the information these representations convey, and specifically to validate the hypothesis that they chiefly encode writing style, we systematically probe these representations through a series of targeted experiments. The results of these experiments suggest that representations learned for the surrogate authorship prediction task are indeed sensitive to writing style. As a consequence, authorship representations may be expected to be robust to certain kinds of data shift, such as topic drift over time. Additionally, our findings may open the door to downstream applications that require stylistic representations, such as style transfer.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"21 1","pages":"1416-1431"},"PeriodicalIF":10.9,"publicationDate":"2023-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139349572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PaniniQA: Enhancing Patient Education Through Interactive Question Answering PaniniQA:通过交互式问题解答加强患者教育
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-07 DOI: 10.1162/tacl_a_00616
Pengshan Cai, Zonghai Yao, Fei Liu, Dakuo Wang, Meghan Reilly, Huixue Zhou, Lingxi Li, Yifan Cao, Alok Kapoor, Adarsha S. Bajracharya, D. Berlowitz, Hongfeng Yu
Abstract A patient portal allows discharged patients to access their personalized discharge instructions in electronic health records (EHRs). However, many patients have difficulty understanding or memorizing their discharge instructions (Zhao et al., 2017). In this paper, we present PaniniQA, a patient-centric interactive question answering system designed to help patients understand their discharge instructions. PaniniQA first identifies important clinical content from patients’ discharge instructions and then formulates patient-specific educational questions. In addition, PaniniQA is also equipped with answer verification functionality to provide timely feedback to correct patients’ misunderstandings. Our comprehensive automatic & human evaluation results demonstrate our PaniniQA is capable of improving patients’ mastery of their medical instructions through effective interactions.1
摘要 患者门户网站允许出院患者访问电子健康记录(EHR)中的个性化出院指导。然而,许多患者很难理解或记住他们的出院指导(Zhao 等人,2017)。在本文中,我们介绍了 PaniniQA,这是一个以患者为中心的交互式问题解答系统,旨在帮助患者理解他们的出院指导。PaniniQA 首先从患者的出院指导中识别出重要的临床内容,然后制定出针对患者的教育问题。此外,PaniniQA 还配备了答案验证功能,可提供及时反馈,纠正患者的误解。我们的自动和人工综合评估结果表明,PaniniQA 能够通过有效的互动提高患者对医嘱的掌握程度1。
{"title":"PaniniQA: Enhancing Patient Education Through Interactive Question Answering","authors":"Pengshan Cai, Zonghai Yao, Fei Liu, Dakuo Wang, Meghan Reilly, Huixue Zhou, Lingxi Li, Yifan Cao, Alok Kapoor, Adarsha S. Bajracharya, D. Berlowitz, Hongfeng Yu","doi":"10.1162/tacl_a_00616","DOIUrl":"https://doi.org/10.1162/tacl_a_00616","url":null,"abstract":"Abstract A patient portal allows discharged patients to access their personalized discharge instructions in electronic health records (EHRs). However, many patients have difficulty understanding or memorizing their discharge instructions (Zhao et al., 2017). In this paper, we present PaniniQA, a patient-centric interactive question answering system designed to help patients understand their discharge instructions. PaniniQA first identifies important clinical content from patients’ discharge instructions and then formulates patient-specific educational questions. In addition, PaniniQA is also equipped with answer verification functionality to provide timely feedback to correct patients’ misunderstandings. Our comprehensive automatic & human evaluation results demonstrate our PaniniQA is capable of improving patients’ mastery of their medical instructions through effective interactions.1","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"2 1","pages":"1518-1536"},"PeriodicalIF":10.9,"publicationDate":"2023-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139351516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning to Paraphrase Sentences to Different Complexity Levels 学习仿写不同复杂程度的句子
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-04 DOI: 10.1162/tacl_a_00606
Alison Chi, Li-Kuang Chen, Yi-Chen Chang, Shu-Hui Lee, Jason J. S. Chang
Abstract While sentence simplification is an active research topic in NLP, its adjacent tasks of sentence complexification and same-level paraphrasing are not. To train models on all three tasks, we present two new unsupervised datasets. We compare these datasets, one labeled by a weak classifier and the other by a rule-based approach, with a single supervised dataset. Using these three datasets for training, we perform extensive experiments on both multitasking and prompting strategies. Compared to other systems trained on unsupervised parallel data, models trained on our weak classifier labeled dataset achieve state-of-the-art performance on the ASSET simplification benchmark. Our models also outperform previous work on sentence-level targeting. Finally, we establish how a handful of Large Language Models perform on these tasks under a zero-shot setting.
摘要 虽然句子简化是 NLP 中一个活跃的研究课题,但与之相邻的句子复杂化和同级意译任务却并不活跃。为了训练这三个任务的模型,我们提出了两个新的无监督数据集。我们将这些数据集(一个由弱分类器标注,另一个由基于规则的方法标注)与单一的有监督数据集进行了比较。利用这三个数据集进行训练,我们对多任务和提示策略进行了广泛的实验。与其他在无监督并行数据上训练的系统相比,在我们的弱分类器标注数据集上训练的模型在 ASSET 简化基准上达到了最先进的性能。我们的模型在句子级目标定位方面的表现也优于之前的工作。最后,我们还确定了一些大型语言模型在零镜头设置下对这些任务的表现。
{"title":"Learning to Paraphrase Sentences to Different Complexity Levels","authors":"Alison Chi, Li-Kuang Chen, Yi-Chen Chang, Shu-Hui Lee, Jason J. S. Chang","doi":"10.1162/tacl_a_00606","DOIUrl":"https://doi.org/10.1162/tacl_a_00606","url":null,"abstract":"Abstract While sentence simplification is an active research topic in NLP, its adjacent tasks of sentence complexification and same-level paraphrasing are not. To train models on all three tasks, we present two new unsupervised datasets. We compare these datasets, one labeled by a weak classifier and the other by a rule-based approach, with a single supervised dataset. Using these three datasets for training, we perform extensive experiments on both multitasking and prompting strategies. Compared to other systems trained on unsupervised parallel data, models trained on our weak classifier labeled dataset achieve state-of-the-art performance on the ASSET simplification benchmark. Our models also outperform previous work on sentence-level targeting. Finally, we establish how a handful of Large Language Models perform on these tasks under a zero-shot setting.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"94 1","pages":"1332-1354"},"PeriodicalIF":10.9,"publicationDate":"2023-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139351690","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collective Human Opinions in Semantic Textual Similarity 语义文本相似度中的人类集体意见
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-01 DOI: 10.1162/tacl_a_00584
Yuxia Wang, Shimin Tao, Ning Xie, Hao Yang, Timothy Baldwin, K. Verspoor
Abstract Despite the subjective nature of semantic textual similarity (STS) and pervasive disagreements in STS annotation, existing benchmarks have used averaged human ratings as gold standard. Averaging masks the true distribution of human opinions on examples of low agreement, and prevents models from capturing the semantic vagueness that the individual ratings represent. In this work, we introduce USTS, the first Uncertainty-aware STS dataset with ∼15,000 Chinese sentence pairs and 150,000 labels, to study collective human opinions in STS. Analysis reveals that neither a scalar nor a single Gaussian fits a set of observed judgments adequately. We further show that current STS models cannot capture the variance caused by human disagreement on individual instances, but rather reflect the predictive confidence over the aggregate dataset.
尽管语义文本相似度(STS)的主观性和STS注释中普遍存在的分歧,但现有的基准都使用平均人类评分作为金标准。平均掩盖了人类对低一致性示例的真实意见分布,并阻止模型捕获单个评级所代表的语义模糊性。在这项工作中,我们引入了USTS,这是第一个具有不确定性感知的STS数据集,包含约15,000个中文句子对和150,000个标签,用于研究STS中的集体人类意见。分析表明,标量和单个高斯分布都不能充分拟合一组观察到的判断。我们进一步表明,目前的STS模型不能捕捉到人类在单个实例上的分歧所引起的方差,而是反映了对总体数据集的预测置信度。
{"title":"Collective Human Opinions in Semantic Textual Similarity","authors":"Yuxia Wang, Shimin Tao, Ning Xie, Hao Yang, Timothy Baldwin, K. Verspoor","doi":"10.1162/tacl_a_00584","DOIUrl":"https://doi.org/10.1162/tacl_a_00584","url":null,"abstract":"Abstract Despite the subjective nature of semantic textual similarity (STS) and pervasive disagreements in STS annotation, existing benchmarks have used averaged human ratings as gold standard. Averaging masks the true distribution of human opinions on examples of low agreement, and prevents models from capturing the semantic vagueness that the individual ratings represent. In this work, we introduce USTS, the first Uncertainty-aware STS dataset with ∼15,000 Chinese sentence pairs and 150,000 labels, to study collective human opinions in STS. Analysis reveals that neither a scalar nor a single Gaussian fits a set of observed judgments adequately. We further show that current STS models cannot capture the variance caused by human disagreement on individual instances, but rather reflect the predictive confidence over the aggregate dataset.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"11 1","pages":"997-1013"},"PeriodicalIF":10.9,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42343606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Time-and-Space-Efficient Weighted Deduction 时间和空间效率加权扣除
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-01 DOI: 10.1162/tacl_a_00588
Jason Eisner
Abstract Many NLP algorithms have been described in terms of deduction systems. Unweighted deduction allows a generic forward-chaining execution strategy. For weighted deduction, however, efficient execution should propagate the weight of each item only after it has converged. This means visiting the items in topologically sorted order (as in dynamic programming). Toposorting is fast on a materialized graph; unfortunately, materializing the graph would take extra space. Is there a generic weighted deduction strategy which, for every acyclic deduction system and every input, uses only a constant factor more time and space than generic unweighted deduction? After reviewing past strategies, we answer this question in the affirmative by combining ideas of Goodman (1999) and Kahn (1962). We also give an extension to cyclic deduction systems, based on Tarjan (1972).
许多NLP算法都是用演绎系统来描述的。非加权扣除允许通用前向链执行策略。然而,对于加权推导,有效的执行应该只在每个项收敛之后才传播其权重。这意味着以拓扑排序的顺序访问项目(如动态规划)。在物化图上拓扑排序速度快;不幸的是,物化图形将占用额外的空间。是否存在一种一般的加权演绎策略,对于每一个非循环演绎系统和每一个输入,只比一般的非加权演绎使用一个常数因子更多的时间和空间?在回顾了过去的策略之后,我们结合Goodman(1999)和Kahn(1962)的观点来肯定地回答这个问题。我们也在Tarjan(1972)的基础上对循环演绎系统进行了扩展。
{"title":"Time-and-Space-Efficient Weighted Deduction","authors":"Jason Eisner","doi":"10.1162/tacl_a_00588","DOIUrl":"https://doi.org/10.1162/tacl_a_00588","url":null,"abstract":"Abstract Many NLP algorithms have been described in terms of deduction systems. Unweighted deduction allows a generic forward-chaining execution strategy. For weighted deduction, however, efficient execution should propagate the weight of each item only after it has converged. This means visiting the items in topologically sorted order (as in dynamic programming). Toposorting is fast on a materialized graph; unfortunately, materializing the graph would take extra space. Is there a generic weighted deduction strategy which, for every acyclic deduction system and every input, uses only a constant factor more time and space than generic unweighted deduction? After reviewing past strategies, we answer this question in the affirmative by combining ideas of Goodman (1999) and Kahn (1962). We also give an extension to cyclic deduction systems, based on Tarjan (1972).","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"11 1","pages":"960-973"},"PeriodicalIF":10.9,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"64440766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi 3 WOZ: A Multilingual, Multi-Domain, Multi-Parallel Dataset for Training and Evaluating Culturally Adapted Task-Oriented Dialog Systems Multi 3 WOZ:用于训练和评估文化适应性任务导向型对话系统的多语言、多领域、多并行数据集
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-07-26 DOI: 10.1162/tacl_a_00609
Songbo Hu, Han Zhou, Mete Hergul, Milan Gritta, Guchun Zhang, Ignacio Iacobacci, Ivan Vulic, A. Korhonen
Abstract Creating high-quality annotated data for task-oriented dialog (ToD) is known to be notoriously difficult, and the challenges are amplified when the goal is to create equitable, culturally adapted, and large-scale ToD datasets for multiple languages. Therefore, the current datasets are still very scarce and suffer from limitations such as translation-based non-native dialogs with translation artefacts, small scale, or lack of cultural adaptation, among others. In this work, we first take stock of the current landscape of multilingual ToD datasets, offering a systematic overview of their properties and limitations. Aiming to reduce all the detected limitations, we then introduce Multi3WOZ, a novel multilingual, multi-domain, multi-parallel ToD dataset. It is large-scale and offers culturally adapted dialogs in 4 languages to enable training and evaluation of multilingual and cross-lingual ToD systems. We describe a complex bottom–up data collection process that yielded the final dataset, and offer the first sets of baseline scores across different ToD-related tasks for future reference, also highlighting its challenging nature.
摘要 众所周知,为面向任务的对话(ToD)创建高质量的注释数据非常困难,而当目标是为多种语言创建公平、文化适应性强和大规模的 ToD 数据集时,挑战就更大了。因此,目前的数据集仍然非常稀缺,而且存在诸多局限性,例如基于翻译的非母语对话存在翻译假象、规模较小或缺乏文化适应性等等。在这项工作中,我们首先对当前的多语言 ToD 数据集进行了评估,系统地概述了这些数据集的特性和局限性。为了减少所有发现的局限性,我们随后介绍了 Multi3WOZ,这是一种新型的多语言、多领域、多并行 ToD 数据集。该数据集规模庞大,提供 4 种语言的文化适应对话,可用于多语言和跨语言 ToD 系统的培训和评估。我们介绍了一个复杂的自下而上的数据收集过程,该过程产生了最终数据集,并提供了不同 ToD 相关任务的首批基线分数供未来参考,同时也强调了该数据集的挑战性。
{"title":"Multi 3 WOZ: A Multilingual, Multi-Domain, Multi-Parallel Dataset for Training and Evaluating Culturally Adapted Task-Oriented Dialog Systems","authors":"Songbo Hu, Han Zhou, Mete Hergul, Milan Gritta, Guchun Zhang, Ignacio Iacobacci, Ivan Vulic, A. Korhonen","doi":"10.1162/tacl_a_00609","DOIUrl":"https://doi.org/10.1162/tacl_a_00609","url":null,"abstract":"Abstract Creating high-quality annotated data for task-oriented dialog (ToD) is known to be notoriously difficult, and the challenges are amplified when the goal is to create equitable, culturally adapted, and large-scale ToD datasets for multiple languages. Therefore, the current datasets are still very scarce and suffer from limitations such as translation-based non-native dialogs with translation artefacts, small scale, or lack of cultural adaptation, among others. In this work, we first take stock of the current landscape of multilingual ToD datasets, offering a systematic overview of their properties and limitations. Aiming to reduce all the detected limitations, we then introduce Multi3WOZ, a novel multilingual, multi-domain, multi-parallel ToD dataset. It is large-scale and offers culturally adapted dialogs in 4 languages to enable training and evaluation of multilingual and cross-lingual ToD systems. We describe a complex bottom–up data collection process that yielded the final dataset, and offer the first sets of baseline scores across different ToD-related tasks for future reference, also highlighting its challenging nature.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"17 1","pages":"1396-1415"},"PeriodicalIF":10.9,"publicationDate":"2023-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139354806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimal Transport Posterior Alignment for Cross-lingual Semantic Parsing 跨语言语义解析的最优传输后验对齐
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-07-09 DOI: 10.1162/tacl_a_00611
Tom Sherborne, Tom Hosking, Mirella Lapata
Abstract Cross-lingual semantic parsing transfers parsing capability from a high-resource language (e.g., English) to low-resource languages with scarce training data. Previous work has primarily considered silver-standard data augmentation or zero-shot methods; exploiting few-shot gold data is comparatively unexplored. We propose a new approach to cross-lingual semantic parsing by explicitly minimizing cross-lingual divergence between probabilistic latent variables using Optimal Transport. We demonstrate how this direct guidance improves parsing from natural languages using fewer examples and less training. We evaluate our method on two datasets, MTOP and MultiATIS++SQL, establishing state-of-the-art results under a few-shot cross-lingual regime. Ablation studies further reveal that our method improves performance even without parallel input translations. In addition, we show that our model better captures cross-lingual structure in the latent space to improve semantic representation similarity.1
摘要 跨语言语义解析将解析能力从高资源语言(如英语)转移到缺乏训练数据的低资源语言。以前的工作主要考虑的是银标准数据扩增或零镜头方法,而利用少镜头黄金数据的方法相对来说还没有被探索过。我们提出了一种新的跨语言语义解析方法,即利用最优传输(Optimal Transport)明确地最小化概率潜变量之间的跨语言分歧。我们展示了这种直接指导是如何利用更少的示例和训练来改进自然语言解析的。我们在 MTOP 和 MultiATIS++SQL 这两个数据集上对我们的方法进行了评估,结果表明,我们的方法在少量跨语言机制下取得了最先进的成果。消融研究进一步表明,即使没有平行输入翻译,我们的方法也能提高性能。此外,我们还证明,我们的模型能更好地捕捉潜在空间中的跨语言结构,从而提高语义表征的相似性。
{"title":"Optimal Transport Posterior Alignment for Cross-lingual Semantic Parsing","authors":"Tom Sherborne, Tom Hosking, Mirella Lapata","doi":"10.1162/tacl_a_00611","DOIUrl":"https://doi.org/10.1162/tacl_a_00611","url":null,"abstract":"Abstract Cross-lingual semantic parsing transfers parsing capability from a high-resource language (e.g., English) to low-resource languages with scarce training data. Previous work has primarily considered silver-standard data augmentation or zero-shot methods; exploiting few-shot gold data is comparatively unexplored. We propose a new approach to cross-lingual semantic parsing by explicitly minimizing cross-lingual divergence between probabilistic latent variables using Optimal Transport. We demonstrate how this direct guidance improves parsing from natural languages using fewer examples and less training. We evaluate our method on two datasets, MTOP and MultiATIS++SQL, establishing state-of-the-art results under a few-shot cross-lingual regime. Ablation studies further reveal that our method improves performance even without parallel input translations. In addition, we show that our model better captures cross-lingual structure in the latent space to improve semantic representation similarity.1","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"43 1","pages":"1432-1450"},"PeriodicalIF":10.9,"publicationDate":"2023-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139361256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Testing the Predictions of Surprisal Theory in 11 Languages 在 11 种语言中测试惊奇理论的预测结果
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-07-07 DOI: 10.1162/tacl_a_00612
Ethan Gotlieb Wilcox, Tiago Pimentel, Clara Meister, Ryan Cotterell, R. Levy
Abstract Surprisal theory posits that less-predictable words should take more time to process, with word predictability quantified as surprisal, i.e., negative log probability in context. While evidence supporting the predictions of surprisal theory has been replicated widely, much of it has focused on a very narrow slice of data: native English speakers reading English texts. Indeed, no comprehensive multilingual analysis exists. We address this gap in the current literature by investigating the relationship between surprisal and reading times in eleven different languages, distributed across five language families. Deriving estimates from language models trained on monolingual and multilingual corpora, we test three predictions associated with surprisal theory: (i) whether surprisal is predictive of reading times, (ii) whether expected surprisal, i.e., contextual entropy, is predictive of reading times, and (iii) whether the linking function between surprisal and reading times is linear. We find that all three predictions are borne out crosslinguistically. By focusing on a more diverse set of languages, we argue that these results offer the most robust link to date between information theory and incremental language processing across languages.
摘要 惊奇理论(surprisal theory)认为,可预测性较低的单词应该需要更多的时间来处理,单词的可预测性量化为惊奇(surprisal),即上下文中的负对数概率。虽然支持惊奇理论预测的证据已被广泛复制,但其中大部分都集中在非常狭窄的数据片段上:以英语为母语的人阅读英语文本。事实上,目前还没有全面的多语言分析。我们通过研究五大语系 11 种不同语言中惊奇和阅读时间之间的关系,填补了目前文献中的这一空白。根据在单语和多语语料库中训练的语言模型得出的估计值,我们检验了与惊奇理论相关的三个预测:(i) 惊奇是否能预测阅读时间;(ii) 预期惊奇(即上下文熵)是否能预测阅读时间;(iii) 惊奇与阅读时间之间的关联函数是否是线性的。我们发现这三个预测在跨语言研究中都得到了证实。通过关注更多样化的语言,我们认为这些结果提供了迄今为止信息理论与跨语言增量语言处理之间最稳健的联系。
{"title":"Testing the Predictions of Surprisal Theory in 11 Languages","authors":"Ethan Gotlieb Wilcox, Tiago Pimentel, Clara Meister, Ryan Cotterell, R. Levy","doi":"10.1162/tacl_a_00612","DOIUrl":"https://doi.org/10.1162/tacl_a_00612","url":null,"abstract":"Abstract Surprisal theory posits that less-predictable words should take more time to process, with word predictability quantified as surprisal, i.e., negative log probability in context. While evidence supporting the predictions of surprisal theory has been replicated widely, much of it has focused on a very narrow slice of data: native English speakers reading English texts. Indeed, no comprehensive multilingual analysis exists. We address this gap in the current literature by investigating the relationship between surprisal and reading times in eleven different languages, distributed across five language families. Deriving estimates from language models trained on monolingual and multilingual corpora, we test three predictions associated with surprisal theory: (i) whether surprisal is predictive of reading times, (ii) whether expected surprisal, i.e., contextual entropy, is predictive of reading times, and (iii) whether the linking function between surprisal and reading times is linear. We find that all three predictions are borne out crosslinguistically. By focusing on a more diverse set of languages, we argue that these results offer the most robust link to date between information theory and incremental language processing across languages.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"5 1","pages":"1451-1470"},"PeriodicalIF":10.9,"publicationDate":"2023-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139362131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rank-Aware Negative Training for Semi-Supervised Text Classification 半监督文本分类的秩感知负训练
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-06-13 DOI: 10.1162/tacl_a_00574
Ahmed Murtadha, Shengfeng Pan, Wen Bo, Jianlin Su, Xinxin Cao, Wenze Zhang, Yunfeng Liu
Abstract Semi-supervised text classification-based paradigms (SSTC) typically employ the spirit of self-training. The key idea is to train a deep classifier on limited labeled texts and then iteratively predict the unlabeled texts as their pseudo-labels for further training. However, the performance is largely affected by the accuracy of pseudo-labels, which may not be significant in real-world scenarios. This paper presents a Rank-aware Negative Training (RNT) framework to address SSTC in learning with noisy label settings. To alleviate the noisy information, we adapt a reasoning with uncertainty-based approach to rank the unlabeled texts based on the evidential support received from the labeled texts. Moreover, we propose the use of negative training to train RNT based on the concept that “the input instance does not belong to the complementary label”. A complementary label is randomly selected from all labels except the label on-target. Intuitively, the probability of a true label serving as a complementary label is low and thus provides less noisy information during the training, resulting in better performance on the test data. Finally, we evaluate the proposed solution on various text classification benchmark datasets. Our extensive experiments show that it consistently overcomes the state-of-the-art alternatives in most scenarios and achieves competitive performance in the others. The code of RNT is publicly available on GitHub.
摘要基于半监督文本分类的范式(SSTC)通常采用自我训练的精神。关键思想是在有限的标记文本上训练一个深度分类器,然后迭代地预测未标记文本作为它们的伪标签,以进行进一步的训练。然而,性能在很大程度上受到伪标签准确性的影响,而伪标签在现实世界中可能并不重要。本文提出了一个秩感知负训练(RNT)框架,以解决在有噪声标签设置的学习中的SSTC问题。为了减轻噪声信息,我们采用了一种基于不确定性的推理方法,根据从标记文本中获得的证据支持对未标记文本进行排序。此外,基于“输入实例不属于互补标签”的概念,我们提出使用负训练来训练RNT。从除目标上的标签外的所有标签中随机选择互补标签。直观地,真实标签充当互补标签的概率较低,因此在训练期间提供较少的噪声信息,从而在测试数据上产生更好的性能。最后,我们在各种文本分类基准数据集上对所提出的解决方案进行了评估。我们的大量实验表明,它在大多数情况下始终克服了最先进的替代方案,并在其他情况下实现了有竞争力的性能。RNT的代码在GitHub上是公开的。
{"title":"Rank-Aware Negative Training for Semi-Supervised Text Classification","authors":"Ahmed Murtadha, Shengfeng Pan, Wen Bo, Jianlin Su, Xinxin Cao, Wenze Zhang, Yunfeng Liu","doi":"10.1162/tacl_a_00574","DOIUrl":"https://doi.org/10.1162/tacl_a_00574","url":null,"abstract":"Abstract Semi-supervised text classification-based paradigms (SSTC) typically employ the spirit of self-training. The key idea is to train a deep classifier on limited labeled texts and then iteratively predict the unlabeled texts as their pseudo-labels for further training. However, the performance is largely affected by the accuracy of pseudo-labels, which may not be significant in real-world scenarios. This paper presents a Rank-aware Negative Training (RNT) framework to address SSTC in learning with noisy label settings. To alleviate the noisy information, we adapt a reasoning with uncertainty-based approach to rank the unlabeled texts based on the evidential support received from the labeled texts. Moreover, we propose the use of negative training to train RNT based on the concept that “the input instance does not belong to the complementary label”. A complementary label is randomly selected from all labels except the label on-target. Intuitively, the probability of a true label serving as a complementary label is low and thus provides less noisy information during the training, resulting in better performance on the test data. Finally, we evaluate the proposed solution on various text classification benchmark datasets. Our extensive experiments show that it consistently overcomes the state-of-the-art alternatives in most scenarios and achieves competitive performance in the others. The code of RNT is publicly available on GitHub.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"11 1","pages":"771-786"},"PeriodicalIF":10.9,"publicationDate":"2023-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49384920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Transactions of the Association for Computational Linguistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1