首页 > 最新文献

Transactions of the Association for Computational Linguistics最新文献

英文 中文
PaniniQA: Enhancing Patient Education Through Interactive Question Answering PaniniQA:通过交互式问题解答加强患者教育
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-07 DOI: 10.1162/tacl_a_00616
Pengshan Cai, Zonghai Yao, Fei Liu, Dakuo Wang, Meghan Reilly, Huixue Zhou, Lingxi Li, Yifan Cao, Alok Kapoor, Adarsha S. Bajracharya, D. Berlowitz, Hongfeng Yu
Abstract A patient portal allows discharged patients to access their personalized discharge instructions in electronic health records (EHRs). However, many patients have difficulty understanding or memorizing their discharge instructions (Zhao et al., 2017). In this paper, we present PaniniQA, a patient-centric interactive question answering system designed to help patients understand their discharge instructions. PaniniQA first identifies important clinical content from patients’ discharge instructions and then formulates patient-specific educational questions. In addition, PaniniQA is also equipped with answer verification functionality to provide timely feedback to correct patients’ misunderstandings. Our comprehensive automatic & human evaluation results demonstrate our PaniniQA is capable of improving patients’ mastery of their medical instructions through effective interactions.1
摘要 患者门户网站允许出院患者访问电子健康记录(EHR)中的个性化出院指导。然而,许多患者很难理解或记住他们的出院指导(Zhao 等人,2017)。在本文中,我们介绍了 PaniniQA,这是一个以患者为中心的交互式问题解答系统,旨在帮助患者理解他们的出院指导。PaniniQA 首先从患者的出院指导中识别出重要的临床内容,然后制定出针对患者的教育问题。此外,PaniniQA 还配备了答案验证功能,可提供及时反馈,纠正患者的误解。我们的自动和人工综合评估结果表明,PaniniQA 能够通过有效的互动提高患者对医嘱的掌握程度1。
{"title":"PaniniQA: Enhancing Patient Education Through Interactive Question Answering","authors":"Pengshan Cai, Zonghai Yao, Fei Liu, Dakuo Wang, Meghan Reilly, Huixue Zhou, Lingxi Li, Yifan Cao, Alok Kapoor, Adarsha S. Bajracharya, D. Berlowitz, Hongfeng Yu","doi":"10.1162/tacl_a_00616","DOIUrl":"https://doi.org/10.1162/tacl_a_00616","url":null,"abstract":"Abstract A patient portal allows discharged patients to access their personalized discharge instructions in electronic health records (EHRs). However, many patients have difficulty understanding or memorizing their discharge instructions (Zhao et al., 2017). In this paper, we present PaniniQA, a patient-centric interactive question answering system designed to help patients understand their discharge instructions. PaniniQA first identifies important clinical content from patients’ discharge instructions and then formulates patient-specific educational questions. In addition, PaniniQA is also equipped with answer verification functionality to provide timely feedback to correct patients’ misunderstandings. Our comprehensive automatic & human evaluation results demonstrate our PaniniQA is capable of improving patients’ mastery of their medical instructions through effective interactions.1","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"2 1","pages":"1518-1536"},"PeriodicalIF":10.9,"publicationDate":"2023-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139351516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning to Paraphrase Sentences to Different Complexity Levels 学习仿写不同复杂程度的句子
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-04 DOI: 10.1162/tacl_a_00606
Alison Chi, Li-Kuang Chen, Yi-Chen Chang, Shu-Hui Lee, Jason J. S. Chang
Abstract While sentence simplification is an active research topic in NLP, its adjacent tasks of sentence complexification and same-level paraphrasing are not. To train models on all three tasks, we present two new unsupervised datasets. We compare these datasets, one labeled by a weak classifier and the other by a rule-based approach, with a single supervised dataset. Using these three datasets for training, we perform extensive experiments on both multitasking and prompting strategies. Compared to other systems trained on unsupervised parallel data, models trained on our weak classifier labeled dataset achieve state-of-the-art performance on the ASSET simplification benchmark. Our models also outperform previous work on sentence-level targeting. Finally, we establish how a handful of Large Language Models perform on these tasks under a zero-shot setting.
摘要 虽然句子简化是 NLP 中一个活跃的研究课题,但与之相邻的句子复杂化和同级意译任务却并不活跃。为了训练这三个任务的模型,我们提出了两个新的无监督数据集。我们将这些数据集(一个由弱分类器标注,另一个由基于规则的方法标注)与单一的有监督数据集进行了比较。利用这三个数据集进行训练,我们对多任务和提示策略进行了广泛的实验。与其他在无监督并行数据上训练的系统相比,在我们的弱分类器标注数据集上训练的模型在 ASSET 简化基准上达到了最先进的性能。我们的模型在句子级目标定位方面的表现也优于之前的工作。最后,我们还确定了一些大型语言模型在零镜头设置下对这些任务的表现。
{"title":"Learning to Paraphrase Sentences to Different Complexity Levels","authors":"Alison Chi, Li-Kuang Chen, Yi-Chen Chang, Shu-Hui Lee, Jason J. S. Chang","doi":"10.1162/tacl_a_00606","DOIUrl":"https://doi.org/10.1162/tacl_a_00606","url":null,"abstract":"Abstract While sentence simplification is an active research topic in NLP, its adjacent tasks of sentence complexification and same-level paraphrasing are not. To train models on all three tasks, we present two new unsupervised datasets. We compare these datasets, one labeled by a weak classifier and the other by a rule-based approach, with a single supervised dataset. Using these three datasets for training, we perform extensive experiments on both multitasking and prompting strategies. Compared to other systems trained on unsupervised parallel data, models trained on our weak classifier labeled dataset achieve state-of-the-art performance on the ASSET simplification benchmark. Our models also outperform previous work on sentence-level targeting. Finally, we establish how a handful of Large Language Models perform on these tasks under a zero-shot setting.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"94 1","pages":"1332-1354"},"PeriodicalIF":10.9,"publicationDate":"2023-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139351690","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collective Human Opinions in Semantic Textual Similarity 语义文本相似度中的人类集体意见
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-01 DOI: 10.1162/tacl_a_00584
Yuxia Wang, Shimin Tao, Ning Xie, Hao Yang, Timothy Baldwin, K. Verspoor
Abstract Despite the subjective nature of semantic textual similarity (STS) and pervasive disagreements in STS annotation, existing benchmarks have used averaged human ratings as gold standard. Averaging masks the true distribution of human opinions on examples of low agreement, and prevents models from capturing the semantic vagueness that the individual ratings represent. In this work, we introduce USTS, the first Uncertainty-aware STS dataset with ∼15,000 Chinese sentence pairs and 150,000 labels, to study collective human opinions in STS. Analysis reveals that neither a scalar nor a single Gaussian fits a set of observed judgments adequately. We further show that current STS models cannot capture the variance caused by human disagreement on individual instances, but rather reflect the predictive confidence over the aggregate dataset.
尽管语义文本相似度(STS)的主观性和STS注释中普遍存在的分歧,但现有的基准都使用平均人类评分作为金标准。平均掩盖了人类对低一致性示例的真实意见分布,并阻止模型捕获单个评级所代表的语义模糊性。在这项工作中,我们引入了USTS,这是第一个具有不确定性感知的STS数据集,包含约15,000个中文句子对和150,000个标签,用于研究STS中的集体人类意见。分析表明,标量和单个高斯分布都不能充分拟合一组观察到的判断。我们进一步表明,目前的STS模型不能捕捉到人类在单个实例上的分歧所引起的方差,而是反映了对总体数据集的预测置信度。
{"title":"Collective Human Opinions in Semantic Textual Similarity","authors":"Yuxia Wang, Shimin Tao, Ning Xie, Hao Yang, Timothy Baldwin, K. Verspoor","doi":"10.1162/tacl_a_00584","DOIUrl":"https://doi.org/10.1162/tacl_a_00584","url":null,"abstract":"Abstract Despite the subjective nature of semantic textual similarity (STS) and pervasive disagreements in STS annotation, existing benchmarks have used averaged human ratings as gold standard. Averaging masks the true distribution of human opinions on examples of low agreement, and prevents models from capturing the semantic vagueness that the individual ratings represent. In this work, we introduce USTS, the first Uncertainty-aware STS dataset with ∼15,000 Chinese sentence pairs and 150,000 labels, to study collective human opinions in STS. Analysis reveals that neither a scalar nor a single Gaussian fits a set of observed judgments adequately. We further show that current STS models cannot capture the variance caused by human disagreement on individual instances, but rather reflect the predictive confidence over the aggregate dataset.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"11 1","pages":"997-1013"},"PeriodicalIF":10.9,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42343606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Time-and-Space-Efficient Weighted Deduction 时间和空间效率加权扣除
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-01 DOI: 10.1162/tacl_a_00588
Jason Eisner
Abstract Many NLP algorithms have been described in terms of deduction systems. Unweighted deduction allows a generic forward-chaining execution strategy. For weighted deduction, however, efficient execution should propagate the weight of each item only after it has converged. This means visiting the items in topologically sorted order (as in dynamic programming). Toposorting is fast on a materialized graph; unfortunately, materializing the graph would take extra space. Is there a generic weighted deduction strategy which, for every acyclic deduction system and every input, uses only a constant factor more time and space than generic unweighted deduction? After reviewing past strategies, we answer this question in the affirmative by combining ideas of Goodman (1999) and Kahn (1962). We also give an extension to cyclic deduction systems, based on Tarjan (1972).
许多NLP算法都是用演绎系统来描述的。非加权扣除允许通用前向链执行策略。然而,对于加权推导,有效的执行应该只在每个项收敛之后才传播其权重。这意味着以拓扑排序的顺序访问项目(如动态规划)。在物化图上拓扑排序速度快;不幸的是,物化图形将占用额外的空间。是否存在一种一般的加权演绎策略,对于每一个非循环演绎系统和每一个输入,只比一般的非加权演绎使用一个常数因子更多的时间和空间?在回顾了过去的策略之后,我们结合Goodman(1999)和Kahn(1962)的观点来肯定地回答这个问题。我们也在Tarjan(1972)的基础上对循环演绎系统进行了扩展。
{"title":"Time-and-Space-Efficient Weighted Deduction","authors":"Jason Eisner","doi":"10.1162/tacl_a_00588","DOIUrl":"https://doi.org/10.1162/tacl_a_00588","url":null,"abstract":"Abstract Many NLP algorithms have been described in terms of deduction systems. Unweighted deduction allows a generic forward-chaining execution strategy. For weighted deduction, however, efficient execution should propagate the weight of each item only after it has converged. This means visiting the items in topologically sorted order (as in dynamic programming). Toposorting is fast on a materialized graph; unfortunately, materializing the graph would take extra space. Is there a generic weighted deduction strategy which, for every acyclic deduction system and every input, uses only a constant factor more time and space than generic unweighted deduction? After reviewing past strategies, we answer this question in the affirmative by combining ideas of Goodman (1999) and Kahn (1962). We also give an extension to cyclic deduction systems, based on Tarjan (1972).","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"11 1","pages":"960-973"},"PeriodicalIF":10.9,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"64440766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi 3 WOZ: A Multilingual, Multi-Domain, Multi-Parallel Dataset for Training and Evaluating Culturally Adapted Task-Oriented Dialog Systems Multi 3 WOZ:用于训练和评估文化适应性任务导向型对话系统的多语言、多领域、多并行数据集
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-07-26 DOI: 10.1162/tacl_a_00609
Songbo Hu, Han Zhou, Mete Hergul, Milan Gritta, Guchun Zhang, Ignacio Iacobacci, Ivan Vulic, A. Korhonen
Abstract Creating high-quality annotated data for task-oriented dialog (ToD) is known to be notoriously difficult, and the challenges are amplified when the goal is to create equitable, culturally adapted, and large-scale ToD datasets for multiple languages. Therefore, the current datasets are still very scarce and suffer from limitations such as translation-based non-native dialogs with translation artefacts, small scale, or lack of cultural adaptation, among others. In this work, we first take stock of the current landscape of multilingual ToD datasets, offering a systematic overview of their properties and limitations. Aiming to reduce all the detected limitations, we then introduce Multi3WOZ, a novel multilingual, multi-domain, multi-parallel ToD dataset. It is large-scale and offers culturally adapted dialogs in 4 languages to enable training and evaluation of multilingual and cross-lingual ToD systems. We describe a complex bottom–up data collection process that yielded the final dataset, and offer the first sets of baseline scores across different ToD-related tasks for future reference, also highlighting its challenging nature.
摘要 众所周知,为面向任务的对话(ToD)创建高质量的注释数据非常困难,而当目标是为多种语言创建公平、文化适应性强和大规模的 ToD 数据集时,挑战就更大了。因此,目前的数据集仍然非常稀缺,而且存在诸多局限性,例如基于翻译的非母语对话存在翻译假象、规模较小或缺乏文化适应性等等。在这项工作中,我们首先对当前的多语言 ToD 数据集进行了评估,系统地概述了这些数据集的特性和局限性。为了减少所有发现的局限性,我们随后介绍了 Multi3WOZ,这是一种新型的多语言、多领域、多并行 ToD 数据集。该数据集规模庞大,提供 4 种语言的文化适应对话,可用于多语言和跨语言 ToD 系统的培训和评估。我们介绍了一个复杂的自下而上的数据收集过程,该过程产生了最终数据集,并提供了不同 ToD 相关任务的首批基线分数供未来参考,同时也强调了该数据集的挑战性。
{"title":"Multi 3 WOZ: A Multilingual, Multi-Domain, Multi-Parallel Dataset for Training and Evaluating Culturally Adapted Task-Oriented Dialog Systems","authors":"Songbo Hu, Han Zhou, Mete Hergul, Milan Gritta, Guchun Zhang, Ignacio Iacobacci, Ivan Vulic, A. Korhonen","doi":"10.1162/tacl_a_00609","DOIUrl":"https://doi.org/10.1162/tacl_a_00609","url":null,"abstract":"Abstract Creating high-quality annotated data for task-oriented dialog (ToD) is known to be notoriously difficult, and the challenges are amplified when the goal is to create equitable, culturally adapted, and large-scale ToD datasets for multiple languages. Therefore, the current datasets are still very scarce and suffer from limitations such as translation-based non-native dialogs with translation artefacts, small scale, or lack of cultural adaptation, among others. In this work, we first take stock of the current landscape of multilingual ToD datasets, offering a systematic overview of their properties and limitations. Aiming to reduce all the detected limitations, we then introduce Multi3WOZ, a novel multilingual, multi-domain, multi-parallel ToD dataset. It is large-scale and offers culturally adapted dialogs in 4 languages to enable training and evaluation of multilingual and cross-lingual ToD systems. We describe a complex bottom–up data collection process that yielded the final dataset, and offer the first sets of baseline scores across different ToD-related tasks for future reference, also highlighting its challenging nature.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"17 1","pages":"1396-1415"},"PeriodicalIF":10.9,"publicationDate":"2023-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139354806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimal Transport Posterior Alignment for Cross-lingual Semantic Parsing 跨语言语义解析的最优传输后验对齐
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-07-09 DOI: 10.1162/tacl_a_00611
Tom Sherborne, Tom Hosking, Mirella Lapata
Abstract Cross-lingual semantic parsing transfers parsing capability from a high-resource language (e.g., English) to low-resource languages with scarce training data. Previous work has primarily considered silver-standard data augmentation or zero-shot methods; exploiting few-shot gold data is comparatively unexplored. We propose a new approach to cross-lingual semantic parsing by explicitly minimizing cross-lingual divergence between probabilistic latent variables using Optimal Transport. We demonstrate how this direct guidance improves parsing from natural languages using fewer examples and less training. We evaluate our method on two datasets, MTOP and MultiATIS++SQL, establishing state-of-the-art results under a few-shot cross-lingual regime. Ablation studies further reveal that our method improves performance even without parallel input translations. In addition, we show that our model better captures cross-lingual structure in the latent space to improve semantic representation similarity.1
摘要 跨语言语义解析将解析能力从高资源语言(如英语)转移到缺乏训练数据的低资源语言。以前的工作主要考虑的是银标准数据扩增或零镜头方法,而利用少镜头黄金数据的方法相对来说还没有被探索过。我们提出了一种新的跨语言语义解析方法,即利用最优传输(Optimal Transport)明确地最小化概率潜变量之间的跨语言分歧。我们展示了这种直接指导是如何利用更少的示例和训练来改进自然语言解析的。我们在 MTOP 和 MultiATIS++SQL 这两个数据集上对我们的方法进行了评估,结果表明,我们的方法在少量跨语言机制下取得了最先进的成果。消融研究进一步表明,即使没有平行输入翻译,我们的方法也能提高性能。此外,我们还证明,我们的模型能更好地捕捉潜在空间中的跨语言结构,从而提高语义表征的相似性。
{"title":"Optimal Transport Posterior Alignment for Cross-lingual Semantic Parsing","authors":"Tom Sherborne, Tom Hosking, Mirella Lapata","doi":"10.1162/tacl_a_00611","DOIUrl":"https://doi.org/10.1162/tacl_a_00611","url":null,"abstract":"Abstract Cross-lingual semantic parsing transfers parsing capability from a high-resource language (e.g., English) to low-resource languages with scarce training data. Previous work has primarily considered silver-standard data augmentation or zero-shot methods; exploiting few-shot gold data is comparatively unexplored. We propose a new approach to cross-lingual semantic parsing by explicitly minimizing cross-lingual divergence between probabilistic latent variables using Optimal Transport. We demonstrate how this direct guidance improves parsing from natural languages using fewer examples and less training. We evaluate our method on two datasets, MTOP and MultiATIS++SQL, establishing state-of-the-art results under a few-shot cross-lingual regime. Ablation studies further reveal that our method improves performance even without parallel input translations. In addition, we show that our model better captures cross-lingual structure in the latent space to improve semantic representation similarity.1","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"43 1","pages":"1432-1450"},"PeriodicalIF":10.9,"publicationDate":"2023-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139361256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Testing the Predictions of Surprisal Theory in 11 Languages 在 11 种语言中测试惊奇理论的预测结果
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-07-07 DOI: 10.1162/tacl_a_00612
Ethan Gotlieb Wilcox, Tiago Pimentel, Clara Meister, Ryan Cotterell, R. Levy
Abstract Surprisal theory posits that less-predictable words should take more time to process, with word predictability quantified as surprisal, i.e., negative log probability in context. While evidence supporting the predictions of surprisal theory has been replicated widely, much of it has focused on a very narrow slice of data: native English speakers reading English texts. Indeed, no comprehensive multilingual analysis exists. We address this gap in the current literature by investigating the relationship between surprisal and reading times in eleven different languages, distributed across five language families. Deriving estimates from language models trained on monolingual and multilingual corpora, we test three predictions associated with surprisal theory: (i) whether surprisal is predictive of reading times, (ii) whether expected surprisal, i.e., contextual entropy, is predictive of reading times, and (iii) whether the linking function between surprisal and reading times is linear. We find that all three predictions are borne out crosslinguistically. By focusing on a more diverse set of languages, we argue that these results offer the most robust link to date between information theory and incremental language processing across languages.
摘要 惊奇理论(surprisal theory)认为,可预测性较低的单词应该需要更多的时间来处理,单词的可预测性量化为惊奇(surprisal),即上下文中的负对数概率。虽然支持惊奇理论预测的证据已被广泛复制,但其中大部分都集中在非常狭窄的数据片段上:以英语为母语的人阅读英语文本。事实上,目前还没有全面的多语言分析。我们通过研究五大语系 11 种不同语言中惊奇和阅读时间之间的关系,填补了目前文献中的这一空白。根据在单语和多语语料库中训练的语言模型得出的估计值,我们检验了与惊奇理论相关的三个预测:(i) 惊奇是否能预测阅读时间;(ii) 预期惊奇(即上下文熵)是否能预测阅读时间;(iii) 惊奇与阅读时间之间的关联函数是否是线性的。我们发现这三个预测在跨语言研究中都得到了证实。通过关注更多样化的语言,我们认为这些结果提供了迄今为止信息理论与跨语言增量语言处理之间最稳健的联系。
{"title":"Testing the Predictions of Surprisal Theory in 11 Languages","authors":"Ethan Gotlieb Wilcox, Tiago Pimentel, Clara Meister, Ryan Cotterell, R. Levy","doi":"10.1162/tacl_a_00612","DOIUrl":"https://doi.org/10.1162/tacl_a_00612","url":null,"abstract":"Abstract Surprisal theory posits that less-predictable words should take more time to process, with word predictability quantified as surprisal, i.e., negative log probability in context. While evidence supporting the predictions of surprisal theory has been replicated widely, much of it has focused on a very narrow slice of data: native English speakers reading English texts. Indeed, no comprehensive multilingual analysis exists. We address this gap in the current literature by investigating the relationship between surprisal and reading times in eleven different languages, distributed across five language families. Deriving estimates from language models trained on monolingual and multilingual corpora, we test three predictions associated with surprisal theory: (i) whether surprisal is predictive of reading times, (ii) whether expected surprisal, i.e., contextual entropy, is predictive of reading times, and (iii) whether the linking function between surprisal and reading times is linear. We find that all three predictions are borne out crosslinguistically. By focusing on a more diverse set of languages, we argue that these results offer the most robust link to date between information theory and incremental language processing across languages.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"5 1","pages":"1451-1470"},"PeriodicalIF":10.9,"publicationDate":"2023-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139362131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rank-Aware Negative Training for Semi-Supervised Text Classification 半监督文本分类的秩感知负训练
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-06-13 DOI: 10.1162/tacl_a_00574
Ahmed Murtadha, Shengfeng Pan, Wen Bo, Jianlin Su, Xinxin Cao, Wenze Zhang, Yunfeng Liu
Abstract Semi-supervised text classification-based paradigms (SSTC) typically employ the spirit of self-training. The key idea is to train a deep classifier on limited labeled texts and then iteratively predict the unlabeled texts as their pseudo-labels for further training. However, the performance is largely affected by the accuracy of pseudo-labels, which may not be significant in real-world scenarios. This paper presents a Rank-aware Negative Training (RNT) framework to address SSTC in learning with noisy label settings. To alleviate the noisy information, we adapt a reasoning with uncertainty-based approach to rank the unlabeled texts based on the evidential support received from the labeled texts. Moreover, we propose the use of negative training to train RNT based on the concept that “the input instance does not belong to the complementary label”. A complementary label is randomly selected from all labels except the label on-target. Intuitively, the probability of a true label serving as a complementary label is low and thus provides less noisy information during the training, resulting in better performance on the test data. Finally, we evaluate the proposed solution on various text classification benchmark datasets. Our extensive experiments show that it consistently overcomes the state-of-the-art alternatives in most scenarios and achieves competitive performance in the others. The code of RNT is publicly available on GitHub.
摘要基于半监督文本分类的范式(SSTC)通常采用自我训练的精神。关键思想是在有限的标记文本上训练一个深度分类器,然后迭代地预测未标记文本作为它们的伪标签,以进行进一步的训练。然而,性能在很大程度上受到伪标签准确性的影响,而伪标签在现实世界中可能并不重要。本文提出了一个秩感知负训练(RNT)框架,以解决在有噪声标签设置的学习中的SSTC问题。为了减轻噪声信息,我们采用了一种基于不确定性的推理方法,根据从标记文本中获得的证据支持对未标记文本进行排序。此外,基于“输入实例不属于互补标签”的概念,我们提出使用负训练来训练RNT。从除目标上的标签外的所有标签中随机选择互补标签。直观地,真实标签充当互补标签的概率较低,因此在训练期间提供较少的噪声信息,从而在测试数据上产生更好的性能。最后,我们在各种文本分类基准数据集上对所提出的解决方案进行了评估。我们的大量实验表明,它在大多数情况下始终克服了最先进的替代方案,并在其他情况下实现了有竞争力的性能。RNT的代码在GitHub上是公开的。
{"title":"Rank-Aware Negative Training for Semi-Supervised Text Classification","authors":"Ahmed Murtadha, Shengfeng Pan, Wen Bo, Jianlin Su, Xinxin Cao, Wenze Zhang, Yunfeng Liu","doi":"10.1162/tacl_a_00574","DOIUrl":"https://doi.org/10.1162/tacl_a_00574","url":null,"abstract":"Abstract Semi-supervised text classification-based paradigms (SSTC) typically employ the spirit of self-training. The key idea is to train a deep classifier on limited labeled texts and then iteratively predict the unlabeled texts as their pseudo-labels for further training. However, the performance is largely affected by the accuracy of pseudo-labels, which may not be significant in real-world scenarios. This paper presents a Rank-aware Negative Training (RNT) framework to address SSTC in learning with noisy label settings. To alleviate the noisy information, we adapt a reasoning with uncertainty-based approach to rank the unlabeled texts based on the evidential support received from the labeled texts. Moreover, we propose the use of negative training to train RNT based on the concept that “the input instance does not belong to the complementary label”. A complementary label is randomly selected from all labels except the label on-target. Intuitively, the probability of a true label serving as a complementary label is low and thus provides less noisy information during the training, resulting in better performance on the test data. Finally, we evaluate the proposed solution on various text classification benchmark datasets. Our extensive experiments show that it consistently overcomes the state-of-the-art alternatives in most scenarios and achieves competitive performance in the others. The code of RNT is publicly available on GitHub.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"11 1","pages":"771-786"},"PeriodicalIF":10.9,"publicationDate":"2023-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49384920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Cross-Linguistic Pressure for Uniform Information Density in Word Order 统一语序信息密度的跨语言压力
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-06-06 DOI: 10.1162/tacl_a_00589
T. Clark, Clara Meister, Tiago Pimentel, Michael Hahn, Ryan Cotterell, Richard Futrell, Roger Levy Mit, E. Zurich, U. Cambridge, Saarland University, UC Irvine
Abstract While natural languages differ widely in both canonical word order and word order flexibility, their word orders still follow shared cross-linguistic statistical patterns, often attributed to functional pressures. In the effort to identify these pressures, prior work has compared real and counterfactual word orders. Yet one functional pressure has been overlooked in such investigations: The uniform information density (UID) hypothesis, which holds that information should be spread evenly throughout an utterance. Here, we ask whether a pressure for UID may have influenced word order patterns cross-linguistically. To this end, we use computational models to test whether real orders lead to greater information uniformity than counterfactual orders. In our empirical study of 10 typologically diverse languages, we find that: (i) among SVO languages, real word orders consistently have greater uniformity than reverse word orders, and (ii) only linguistically implausible counterfactual orders consistently exceed the uniformity of real orders. These findings are compatible with a pressure for information uniformity in the development and usage of natural languages.1
摘要尽管自然语言在规范语序和语序灵活性方面存在很大差异,但它们的语序仍然遵循共同的跨语言统计模式,这通常归因于功能压力。为了识别这些压力,先前的工作对真实语序和反事实语序进行了比较。然而,在这些研究中,有一种功能压力被忽视了:统一信息密度假说,认为信息应该在整个话语中均匀分布。在这里,我们要问UID的压力是否影响了语序模式的跨语言性。为此,我们使用计算模型来测试真实订单是否比反事实订单带来更大的信息一致性。在我们对10种类型多样的语言的实证研究中,我们发现:(i)在SVO语言中,真实语序始终比反向语序具有更大的一致性,以及(ii)只有在语言上不可信的反事实语序始终超过真实语序的一致性。这些发现与自然语言发展和使用中信息统一的压力相一致。1
{"title":"A Cross-Linguistic Pressure for Uniform Information Density in Word Order","authors":"T. Clark, Clara Meister, Tiago Pimentel, Michael Hahn, Ryan Cotterell, Richard Futrell, Roger Levy Mit, E. Zurich, U. Cambridge, Saarland University, UC Irvine","doi":"10.1162/tacl_a_00589","DOIUrl":"https://doi.org/10.1162/tacl_a_00589","url":null,"abstract":"Abstract While natural languages differ widely in both canonical word order and word order flexibility, their word orders still follow shared cross-linguistic statistical patterns, often attributed to functional pressures. In the effort to identify these pressures, prior work has compared real and counterfactual word orders. Yet one functional pressure has been overlooked in such investigations: The uniform information density (UID) hypothesis, which holds that information should be spread evenly throughout an utterance. Here, we ask whether a pressure for UID may have influenced word order patterns cross-linguistically. To this end, we use computational models to test whether real orders lead to greater information uniformity than counterfactual orders. In our empirical study of 10 typologically diverse languages, we find that: (i) among SVO languages, real word orders consistently have greater uniformity than reverse word orders, and (ii) only linguistically implausible counterfactual orders consistently exceed the uniformity of real orders. These findings are compatible with a pressure for information uniformity in the development and usage of natural languages.1","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"11 1","pages":"1048-1065"},"PeriodicalIF":10.9,"publicationDate":"2023-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48703963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Supervised Gradual Machine Learning for Aspect-Term Sentiment Analysis 用于方面项情感分析的监督渐进机器学习
IF 10.9 1区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-06-01 DOI: 10.1162/tacl_a_00571
Yanyan Wang, Qun Chen, Murtadha Ahmed, Zhaoqiang Chen, Jing Su, Wei Pan, Zhanhuai Li
Recent work has shown that Aspect-Term Sentiment Analysis (ATSA) can be effectively performed by Gradual Machine Learning (GML). However, the performance of the current unsupervised solution is limited by inaccurate and insufficient knowledge conveyance. In this paper, we propose a supervised GML approach for ATSA, which can effectively exploit labeled training data to improve knowledge conveyance. It leverages binary polarity relations between instances, which can be either similar or opposite, to enable supervised knowledge conveyance. Besides the explicit polarity relations indicated by discourse structures, it also separately supervises a polarity classification DNN and a binary Siamese network to extract implicit polarity relations. The proposed approach fulfills knowledge conveyance by modeling detected relations as binary features in a factor graph. Our extensive experiments on real benchmark data show that it achieves the state-of-the-art performance across all the test workloads. Our work demonstrates clearly that, in collaboration with DNN for feature extraction, GML outperforms pure DNN solutions.
最近的研究表明,方面术语情感分析(ATSA)可以通过渐进机器学习(GML)有效地进行。然而,当前无监督解决方案的性能受到知识传递不准确和不充分的限制。在本文中,我们提出了一种用于ATSA的监督GML方法,该方法可以有效地利用标记的训练数据来改进知识传递。它利用实例之间的二极性关系,可以是相似的,也可以是相反的,以实现有监督的知识传递。除了话语结构所指示的显性极性关系外,它还分别监督极性分类DNN和二元暹罗网络来提取隐性极性关系。所提出的方法通过将检测到的关系建模为因子图中的二进制特征来实现知识传递。我们在真实基准数据上进行的大量实验表明,它在所有测试工作负载中都实现了最先进的性能。我们的工作清楚地表明,与DNN合作进行特征提取,GML优于纯DNN解决方案。
{"title":"Supervised Gradual Machine Learning for Aspect-Term Sentiment Analysis","authors":"Yanyan Wang, Qun Chen, Murtadha Ahmed, Zhaoqiang Chen, Jing Su, Wei Pan, Zhanhuai Li","doi":"10.1162/tacl_a_00571","DOIUrl":"https://doi.org/10.1162/tacl_a_00571","url":null,"abstract":"Recent work has shown that Aspect-Term Sentiment Analysis (ATSA) can be effectively performed by Gradual Machine Learning (GML). However, the performance of the current unsupervised solution is limited by inaccurate and insufficient knowledge conveyance. In this paper, we propose a supervised GML approach for ATSA, which can effectively exploit labeled training data to improve knowledge conveyance. It leverages binary polarity relations between instances, which can be either similar or opposite, to enable supervised knowledge conveyance. Besides the explicit polarity relations indicated by discourse structures, it also separately supervises a polarity classification DNN and a binary Siamese network to extract implicit polarity relations. The proposed approach fulfills knowledge conveyance by modeling detected relations as binary features in a factor graph. Our extensive experiments on real benchmark data show that it achieves the state-of-the-art performance across all the test workloads. Our work demonstrates clearly that, in collaboration with DNN for feature extraction, GML outperforms pure DNN solutions.","PeriodicalId":33559,"journal":{"name":"Transactions of the Association for Computational Linguistics","volume":"11 1","pages":"723-739"},"PeriodicalIF":10.9,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49253493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Transactions of the Association for Computational Linguistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1