首页 > 最新文献

Computational Linguistics最新文献

英文 中文
Curing the SICK and Other NLI Maladies 治愈SICK和其他NLI Maladies
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-10-12 DOI: 10.1162/coli_a_00465
A. Kalouli, Hai Hu, Alexander F. Webb, Larry Moss, Valeria C V de Paiva
Against the backdrop of the ever-improving Natural Language Inference (NLI) models, recent efforts have focused on the suitability of the current NLI datasets and on the feasibility of the NLI task as it is currently approached. Many of the recent studies have exposed the inherent human disagreements of the inference task and have proposed a shift from categorical labels to human subjective probability assessments, capturing human uncertainty. In this work, we show how neither the current task formulation nor the proposed uncertainty gradient are entirely suitable for solving the NLI challenges. Instead, we propose an ordered sense space annotation, which distinguishes between logical and common-sense inference. One end of the space captures non-sensical inferences, while the other end represents strictly logical scenarios. In the middle of the space, we find a continuum of common-sense, namely, the subjective and graded opinion of a “person on the street.” To arrive at the proposed annotation scheme, we perform a careful investigation of the SICK corpus and we create a taxonomy of annotation issues and guidelines. We re-annotate the corpus with the proposed annotation scheme, utilizing four symbolic inference systems, and then perform a thorough evaluation of the scheme by fine-tuning and testing commonly used pre-trained language models on the re-annotated SICK within various settings. We also pioneer a crowd annotation of a small portion of the MultiNLI corpus, showcasing that it is possible to adapt our scheme for annotation by non-experts on another NLI corpus. Our work shows the efficiency and benefits of the proposed mechanism and opens the way for a careful NLI task refinement.
在不断改进的自然语言推理(NLI)模型的背景下,最近的努力集中在当前NLI数据集的适用性和NLI任务的可行性上,因为它目前正在接近。最近的许多研究都揭示了人类在推理任务中固有的分歧,并提出了从分类标签到人类主观概率评估的转变,以捕捉人类的不确定性。在这项工作中,我们展示了当前的任务公式和提出的不确定性梯度如何都不完全适合解决NLI挑战。相反,我们提出了一个有序的感觉空间注释,它区分逻辑推理和常识推理。空间的一端捕获无意义的推理,而另一端则表示严格的逻辑场景。在空间的中间,我们发现了一个常识的连续体,即“街上的人”的主观和分级的意见。为了得到建议的注释方案,我们对SICK语料库进行了仔细的调查,并创建了注释问题和指南的分类法。我们利用四种符号推理系统,用提出的标注方案对语料库进行重新标注,然后在不同设置下对重新标注的SICK上常用的预训练语言模型进行微调和测试,对该方案进行全面评估。我们还率先对多语料库的一小部分进行了群体标注,这表明我们的方案可以适用于非专家对另一个多语料库的标注。我们的工作显示了所提出的机制的效率和好处,并为仔细的NLI任务改进开辟了道路。
{"title":"Curing the SICK and Other NLI Maladies","authors":"A. Kalouli, Hai Hu, Alexander F. Webb, Larry Moss, Valeria C V de Paiva","doi":"10.1162/coli_a_00465","DOIUrl":"https://doi.org/10.1162/coli_a_00465","url":null,"abstract":"Against the backdrop of the ever-improving Natural Language Inference (NLI) models, recent efforts have focused on the suitability of the current NLI datasets and on the feasibility of the NLI task as it is currently approached. Many of the recent studies have exposed the inherent human disagreements of the inference task and have proposed a shift from categorical labels to human subjective probability assessments, capturing human uncertainty. In this work, we show how neither the current task formulation nor the proposed uncertainty gradient are entirely suitable for solving the NLI challenges. Instead, we propose an ordered sense space annotation, which distinguishes between logical and common-sense inference. One end of the space captures non-sensical inferences, while the other end represents strictly logical scenarios. In the middle of the space, we find a continuum of common-sense, namely, the subjective and graded opinion of a “person on the street.” To arrive at the proposed annotation scheme, we perform a careful investigation of the SICK corpus and we create a taxonomy of annotation issues and guidelines. We re-annotate the corpus with the proposed annotation scheme, utilizing four symbolic inference systems, and then perform a thorough evaluation of the scheme by fine-tuning and testing commonly used pre-trained language models on the re-annotated SICK within various settings. We also pioneer a crowd annotation of a small portion of the MultiNLI corpus, showcasing that it is possible to adapt our scheme for annotation by non-experts on another NLI corpus. Our work shows the efficiency and benefits of the proposed mechanism and opens the way for a careful NLI task refinement.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48889363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Martha Evens, Brief Autobiography Martha Evens,简介
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-08-29 DOI: 10.1162/coli_a_00452
M. Evens
{"title":"Martha Evens, Brief Autobiography","authors":"M. Evens","doi":"10.1162/coli_a_00452","DOIUrl":"https://doi.org/10.1162/coli_a_00452","url":null,"abstract":"","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49111307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Martha Palmer and Barbara Di Eugenio Interview Martha Evens 玛莎·帕尔默和芭芭拉·迪·尤金尼奥采访玛莎·埃文斯
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-08-29 DOI: 10.1162/coli_a_00453
Martha Evens
strategies and student behaviors, including differences between face-to-face and computer-mediated tutoring sessions; the usage of hinting and of analogies on the part of the tutor; taking initiative on the part of the students; and several domain-based teaching techniques, for example, at which level of knowledge to teach. All of these strategies were implemented, and several were evaluated in careful experiments. CIRCSIM-Tutor was shown to engender significant learning gains, and was used in actual classes, which is even more striking since the NLP technologies available at the time were severely limited. For further details, please see Di Eugenio et al. 3 ]
策略和学生行为,包括面对面和电脑辅导课程之间的差异;教师暗示和类比的用法;发挥学生的主动性;还有一些基于领域的教学技巧,例如,在哪个层次的知识上进行教学。所有这些策略都被实施了,其中一些在仔细的实验中进行了评估。CIRCSIM-Tutor被证明可以产生显著的学习收益,并在实际课堂中使用,这更加引人注目,因为当时可用的NLP技术严重有限。欲了解更多详情,请参见Di Eugenio等。[3]
{"title":"Martha Palmer and Barbara Di Eugenio Interview Martha Evens","authors":"Martha Evens","doi":"10.1162/coli_a_00453","DOIUrl":"https://doi.org/10.1162/coli_a_00453","url":null,"abstract":"strategies and student behaviors, including differences between face-to-face and computer-mediated tutoring sessions; the usage of hinting and of analogies on the part of the tutor; taking initiative on the part of the students; and several domain-based teaching techniques, for example, at which level of knowledge to teach. All of these strategies were implemented, and several were evaluated in careful experiments. CIRCSIM-Tutor was shown to engender significant learning gains, and was used in actual classes, which is even more striking since the NLP technologies available at the time were severely limited. For further details, please see Di Eugenio et al. 3 ]","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48552327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Explainable Natural Language Processing 可解释的自然语言处理
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-08-26 DOI: 10.1162/coli_r_00460
G. Chrysostomou
Explainable Natural Language Processing (NLP) is an emerging field, which has received significant attention from the NLP community in the last few years. At its core is the need to explain the predictions of machine learning models, now more frequently deployed and used in sensitive areas such as healthcare and law. The rapid developments in the area of explainable NLP have led to somewhat disconnected groups of studies working on these areas. This disconnect results in researchers adopting various definitions for similar problems, while also in certain cases enabling the re-creation of previous research, highlighting the need for a unified framework for explainable NLP. Written by Anders Søgaard, this book provides the author’s convincing view of how we should first define explanations, and, secondly, how we should categorize explanations and the approaches that generate them, creating first and foremost a taxonomy and a unified framework for explainable NLP. As per the author, this will make it easier to relate studies and explanation methodologies in this field, with the aim of accelerating research. It is a brilliant book for both researchers starting to explore explainable NLP, but also for researchers with experience in this area, as it provides a holistic up-to-date view of the explainable NLP at the local and global level. The author conveniently and logically presents each chapter as a “problem” of explainable NLP, as such providing also a taxonomy of explainable NLP problem areas and current approaches to tackle them. Under each chapter, explanation methods are described in detail, beginning initially with “foundational” approaches (e.g., vanilla gradients) and building toward more complex ones (e.g., integrated gradients). To complement the theory and make this into a complete guide to explainable NLP, the author also describes evaluation approaches and provides a list of datasets and code repositories. As such, although the book requires some basic knowledge of NLP and Machine Learning to get started, it is nevertheless accessible to a large audience. This book is organized into thirteen chapters. In the first chapter the author introduces the problems associated with previously proposed taxonomies for explainable NLP. Chapter 2 follows by introducing popular machine learning architectures used in NLP, while also introducing the explanation taxonomy proposed in the book. Chapters
可解释自然语言处理(NLP)是一个新兴的领域,近年来受到了NLP界的极大关注。其核心是需要解释机器学习模型的预测,机器学习模型现在更频繁地部署和使用在医疗保健和法律等敏感领域。可解释NLP领域的快速发展导致了在这些领域工作的研究小组有些脱节。这种脱节导致研究人员对类似问题采用了不同的定义,同时在某些情况下也允许重新创建以前的研究,这突出了对可解释的NLP的统一框架的必要性。这本书由Anders Søgaard撰写,提供了作者令人信服的观点,即我们应该如何首先定义解释,其次,我们应该如何对解释进行分类以及产生解释的方法,首先为可解释的NLP创建了一个分类法和统一框架。根据作者的说法,这将使该领域的研究和解释方法更容易联系起来,目的是加快研究。对于两位开始探索可解释NLP的研究人员来说,这本书都是一本精彩的书,对于在这一领域有经验的研究人员也是如此,因为它在地方和全球层面上提供了可解释的NLP的全面最新观点。作者方便而合乎逻辑地将每一章都作为可解释的NLP的“问题”来呈现,因此还提供了可解释NLP问题领域的分类法和当前解决这些问题的方法。在每一章中,都详细描述了解释方法,从最初的“基础”方法(如香草梯度)开始,到更复杂的方法(如综合梯度)。为了补充这一理论,并使其成为可解释NLP的完整指南,作者还描述了评估方法,并提供了数据集和代码库列表。因此,尽管这本书需要一些NLP和机器学习的基本知识才能入门,但它仍然可以被大量读者阅读。这本书共分为十三章。在第一章中,作者介绍了与先前提出的可解释NLP分类法相关的问题。第2章介绍了NLP中常用的机器学习体系结构,同时也介绍了书中提出的解释分类法。章节
{"title":"Explainable Natural Language Processing","authors":"G. Chrysostomou","doi":"10.1162/coli_r_00460","DOIUrl":"https://doi.org/10.1162/coli_r_00460","url":null,"abstract":"Explainable Natural Language Processing (NLP) is an emerging field, which has received significant attention from the NLP community in the last few years. At its core is the need to explain the predictions of machine learning models, now more frequently deployed and used in sensitive areas such as healthcare and law. The rapid developments in the area of explainable NLP have led to somewhat disconnected groups of studies working on these areas. This disconnect results in researchers adopting various definitions for similar problems, while also in certain cases enabling the re-creation of previous research, highlighting the need for a unified framework for explainable NLP. Written by Anders Søgaard, this book provides the author’s convincing view of how we should first define explanations, and, secondly, how we should categorize explanations and the approaches that generate them, creating first and foremost a taxonomy and a unified framework for explainable NLP. As per the author, this will make it easier to relate studies and explanation methodologies in this field, with the aim of accelerating research. It is a brilliant book for both researchers starting to explore explainable NLP, but also for researchers with experience in this area, as it provides a holistic up-to-date view of the explainable NLP at the local and global level. The author conveniently and logically presents each chapter as a “problem” of explainable NLP, as such providing also a taxonomy of explainable NLP problem areas and current approaches to tackle them. Under each chapter, explanation methods are described in detail, beginning initially with “foundational” approaches (e.g., vanilla gradients) and building toward more complex ones (e.g., integrated gradients). To complement the theory and make this into a complete guide to explainable NLP, the author also describes evaluation approaches and provides a list of datasets and code repositories. As such, although the book requires some basic knowledge of NLP and Machine Learning to get started, it is nevertheless accessible to a large audience. This book is organized into thirteen chapters. In the first chapter the author introduces the problems associated with previously proposed taxonomies for explainable NLP. Chapter 2 follows by introducing popular machine learning architectures used in NLP, while also introducing the explanation taxonomy proposed in the book. Chapters","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47611632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
How Much Does Lookahead Matter for Disambiguation? Partial Arabic Diacritization Case Study 展望对消除歧义有多重要?部分阿拉伯语转调个案研究
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-08-24 DOI: 10.1162/coli_a_00456
Saeed Esmail, Kfir Bar, N. Dershowitz
Abstract We suggest a model for partial diacritization of deep orthographies. We focus on Arabic, where the optional indication of selected vowels by means of diacritics can resolve ambiguity and improve readability. Our partial diacritizer restores short vowels only when they contribute to the ease of understandability during reading a given running text. The idea is to identify those uncertainties of absent vowels that require the reader to look ahead to disambiguate. To achieve this, two independent neural networks are used for predicting diacritics, one that takes the entire sentence as input and another that considers only the text that has been read thus far. Partial diacritization is then determined by retaining precisely those vowels on which the two networks disagree, preferring the reading based on consideration of the whole sentence over the more naïve reading-order diacritization. For evaluation, we prepared a new dataset of Arabic texts with both full and partial vowelization. In addition to facilitating readability, we find that our partial diacritizer improves translation quality compared either to their total absence or to random selection. Lastly, we study the benefit of knowing the text that follows the word in focus toward the restoration of short vowels during reading, and we measure the degree to which lookahead contributes to resolving ambiguities encountered while reading. L’Herbelot had asserted, that the most ancient Korans, written in the Cufic character, had no vowel points; and that these were first invented by Jahia–ben Jamer, who died in the 127th year of the Hegira. “Toderini’s History of Turkish Literature,” Analytical Review (1789)
摘要我们提出了一个深度正字法的部分变音模型。我们关注阿拉伯语,在阿拉伯语中,通过变音符号对所选元音的可选指示可以解决歧义并提高可读性。只有当短元音有助于在阅读给定的运行文本时易于理解时,我们的部分变音器才能恢复短元音。这个想法是为了识别缺失元音的不确定性,这些不确定性需要读者向前看以消除歧义。为了实现这一点,使用了两个独立的神经网络来预测变音符号,一个将整个句子作为输入,另一个只考虑迄今为止阅读过的文本。然后,通过准确地保留两个网络不一致的元音来确定部分变音,更喜欢基于整个句子的阅读,而不是更天真的阅读顺序变音。为了进行评估,我们准备了一个新的阿拉伯语文本数据集,包括完整和部分元音。除了提高可读性外,我们还发现,与完全不存在或随机选择相比,我们的部分变音器提高了翻译质量。最后,我们研究了在阅读过程中,了解单词后面的文本对恢复短元音的好处,并衡量了前瞻性在解决阅读中遇到的歧义方面的作用。L’Herbelot断言,最古老的《古兰经》是用库菲克文字写成的,没有元音点;这些最早是由贾希亚-本·贾米尔发明的,他死于赫吉拉127年。《托代里尼的土耳其文学史》,《分析评论》(1789)
{"title":"How Much Does Lookahead Matter for Disambiguation? Partial Arabic Diacritization Case Study","authors":"Saeed Esmail, Kfir Bar, N. Dershowitz","doi":"10.1162/coli_a_00456","DOIUrl":"https://doi.org/10.1162/coli_a_00456","url":null,"abstract":"Abstract We suggest a model for partial diacritization of deep orthographies. We focus on Arabic, where the optional indication of selected vowels by means of diacritics can resolve ambiguity and improve readability. Our partial diacritizer restores short vowels only when they contribute to the ease of understandability during reading a given running text. The idea is to identify those uncertainties of absent vowels that require the reader to look ahead to disambiguate. To achieve this, two independent neural networks are used for predicting diacritics, one that takes the entire sentence as input and another that considers only the text that has been read thus far. Partial diacritization is then determined by retaining precisely those vowels on which the two networks disagree, preferring the reading based on consideration of the whole sentence over the more naïve reading-order diacritization. For evaluation, we prepared a new dataset of Arabic texts with both full and partial vowelization. In addition to facilitating readability, we find that our partial diacritizer improves translation quality compared either to their total absence or to random selection. Lastly, we study the benefit of knowing the text that follows the word in focus toward the restoration of short vowels during reading, and we measure the degree to which lookahead contributes to resolving ambiguities encountered while reading. L’Herbelot had asserted, that the most ancient Korans, written in the Cufic character, had no vowel points; and that these were first invented by Jahia–ben Jamer, who died in the 127th year of the Hegira. “Toderini’s History of Turkish Literature,” Analytical Review (1789)","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45679521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
The Analysis of Synonymy and Antonymy in Discourse Relations: An Interpretable Modeling Approach 语篇关系中的同义词和反义词分析:一种可解释的建模方法
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-08-09 DOI: 10.1162/coli_a_00477
Assela Reig-Alamillo, David Torres-Moreno, Eliseo Morales-González, Mauricio Toledo-Acosta, Antoine Taroni, Jorge Hermosillo Valadez
The idea that discourse relations are interpreted both by explicit content and by shared knowledge between producer and interpreter is pervasive in discourse and linguistic studies. How much weight should be ascribed in this process to the lexical semantics of the arguments is, however, uncertain. We propose a computational approach to analyze contrast and concession relations in the PDTB corpus. Our work sheds light on the question of how much lexical relations contribute to the signaling of such explicit and implicit relations, as well as on the contribution of different parts of speech to these semantic relations. This study contributes to bridging the gap between corpus and computational linguistics by proposing transparent and explainable computational models of discourse relations based on the synonymy and antonymy of their arguments.
话语关系既由显性内容解释,又由生产者和解释者之间的共享知识解释,这一观点在话语和语言学研究中普遍存在。然而,在这个过程中,应该给参数的词汇语义赋予多少权重是不确定的。我们提出了一种计算方法来分析PDTB语料库中的对比和让步关系。我们的工作揭示了词汇关系对这种显性和隐性关系的信号传导有多大贡献的问题,以及不同词性对这些语义关系的贡献。本研究通过提出基于论点的同义词和反义词的透明和可解释的话语关系计算模型,弥合了语料库和计算语言学之间的差距。
{"title":"The Analysis of Synonymy and Antonymy in Discourse Relations: An Interpretable Modeling Approach","authors":"Assela Reig-Alamillo, David Torres-Moreno, Eliseo Morales-González, Mauricio Toledo-Acosta, Antoine Taroni, Jorge Hermosillo Valadez","doi":"10.1162/coli_a_00477","DOIUrl":"https://doi.org/10.1162/coli_a_00477","url":null,"abstract":"The idea that discourse relations are interpreted both by explicit content and by shared knowledge between producer and interpreter is pervasive in discourse and linguistic studies. How much weight should be ascribed in this process to the lexical semantics of the arguments is, however, uncertain. We propose a computational approach to analyze contrast and concession relations in the PDTB corpus. Our work sheds light on the question of how much lexical relations contribute to the signaling of such explicit and implicit relations, as well as on the contribution of different parts of speech to these semantic relations. This study contributes to bridging the gap between corpus and computational linguistics by proposing transparent and explainable computational models of discourse relations based on the synonymy and antonymy of their arguments.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43896109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Information Theory–based Compositional Distributional Semantics 基于信息论的组合分布语义
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-08-05 DOI: 10.1162/_
Enrique Amigó, Alejandro Ariza-Casabona, V. Fresno, M. A. Martí
Abstract In the context of text representation, Compositional Distributional Semantics models aim to fuse the Distributional Hypothesis and the Principle of Compositionality. Text embedding is based on co-ocurrence distributions and the representations are in turn combined by compositional functions taking into account the text structure. However, the theoretical basis of compositional functions is still an open issue. In this article we define and study the notion of Information Theory–based Compositional Distributional Semantics (ICDS): (i) We first establish formal properties for embedding, composition, and similarity functions based on Shannon’s Information Theory; (ii) we analyze the existing approaches under this prism, checking whether or not they comply with the established desirable properties; (iii) we propose two parameterizable composition and similarity functions that generalize traditional approaches while fulfilling the formal properties; and finally (iv) we perform an empirical study on several textual similarity datasets that include sentences with a high and low lexical overlap, and on the similarity between words and their description. Our theoretical analysis and empirical results show that fulfilling formal properties affects positively the accuracy of text representation models in terms of correspondence (isometry) between the embedding and meaning spaces.
摘要在文本表征的背景下,组合分布语义模型旨在融合分布假设和组合原则。文本嵌入是基于共存在分布的,而表示又通过考虑文本结构的组合函数进行组合。然而,组成函数的理论基础仍然是一个悬而未决的问题。在本文中,我们定义并研究了基于信息论的组合分布语义(ICDS)的概念:(i)我们首先基于香农信息论建立了嵌入、组合和相似函数的形式性质;(ii)我们在这个棱镜下分析现有的方法,检查它们是否符合既定的理想性质;(iii)我们提出了两个可参数化的组合和相似函数,它们在满足形式性质的同时推广了传统方法;最后(iv)我们对几个文本相似性数据集进行了实证研究,这些数据集包括具有高和低词汇重叠的句子,以及单词之间的相似性及其描述。我们的理论分析和实证结果表明,在嵌入空间和意义空间之间的对应(等距)方面,实现形式属性对文本表示模型的准确性产生了积极影响。
{"title":"Information Theory–based Compositional Distributional Semantics","authors":"Enrique Amigó, Alejandro Ariza-Casabona, V. Fresno, M. A. Martí","doi":"10.1162/_","DOIUrl":"https://doi.org/10.1162/_","url":null,"abstract":"Abstract In the context of text representation, Compositional Distributional Semantics models aim to fuse the Distributional Hypothesis and the Principle of Compositionality. Text embedding is based on co-ocurrence distributions and the representations are in turn combined by compositional functions taking into account the text structure. However, the theoretical basis of compositional functions is still an open issue. In this article we define and study the notion of Information Theory–based Compositional Distributional Semantics (ICDS): (i) We first establish formal properties for embedding, composition, and similarity functions based on Shannon’s Information Theory; (ii) we analyze the existing approaches under this prism, checking whether or not they comply with the established desirable properties; (iii) we propose two parameterizable composition and similarity functions that generalize traditional approaches while fulfilling the formal properties; and finally (iv) we perform an empirical study on several textual similarity datasets that include sentences with a high and low lexical overlap, and on the similarity between words and their description. Our theoretical analysis and empirical results show that fulfilling formal properties affects positively the accuracy of text representation models in terms of correspondence (isometry) between the embedding and meaning spaces.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47598974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effective Approaches to Neural Query Language Identification 神经查询语言识别的有效方法
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-07-18 DOI: 10.1162/coli_a_00451
Xingzhang Ren, Baosong Yang, Dayiheng Liu, Haibo Zhang, Xiaoyu Lv, Liang Yao, Jun Xie
Abstract Query language identification (Q-LID) plays a crucial role in a cross-lingual search engine. There exist two main challenges in Q-LID: (1) insufficient contextual information in queries for disambiguation; and (2) the lack of query-style training examples for low-resource languages. In this article, we propose a neural Q-LID model by alleviating the above problems from both model architecture and data augmentation perspectives. Concretely, we build our model upon the advanced Transformer model. In order to enhance the discrimination of queries, a variety of external features (e.g., character, word, as well as script) are fed into the model and fused by a multi-scale attention mechanism. Moreover, to remedy the low resource challenge in this task, a novel machine translation–based strategy is proposed to automatically generate synthetic query-style data for low-resource languages. We contribute the first Q-LID test set called QID-21, which consists of search queries in 21 languages. Experimental results reveal that our model yields better classification accuracy than strong baselines and existing LID systems on both query and traditional LID tasks.1
摘要查询语言识别在跨语言搜索引擎中起着至关重要的作用。Q-LID存在两个主要问题:(1)查询中上下文信息不足,无法消除歧义;(2)缺乏针对低资源语言的查询式训练示例。在本文中,我们提出了一个神经Q-LID模型,从模型架构和数据增强的角度缓解了上述问题。具体地说,我们在高级Transformer模型的基础上构建我们的模型。为了提高查询的辨别能力,将各种外部特征(如字符、单词和脚本)输入到模型中,并通过多尺度注意机制进行融合。此外,为了解决该任务中的低资源挑战,提出了一种新的基于机器翻译的策略来自动生成低资源语言的综合查询式数据。我们贡献了第一个Q-LID测试集,称为QID-21,它由21种语言的搜索查询组成。实验结果表明,在查询和传统LID任务上,我们的模型比强基线和现有LID系统具有更好的分类精度
{"title":"Effective Approaches to Neural Query Language Identification","authors":"Xingzhang Ren, Baosong Yang, Dayiheng Liu, Haibo Zhang, Xiaoyu Lv, Liang Yao, Jun Xie","doi":"10.1162/coli_a_00451","DOIUrl":"https://doi.org/10.1162/coli_a_00451","url":null,"abstract":"Abstract Query language identification (Q-LID) plays a crucial role in a cross-lingual search engine. There exist two main challenges in Q-LID: (1) insufficient contextual information in queries for disambiguation; and (2) the lack of query-style training examples for low-resource languages. In this article, we propose a neural Q-LID model by alleviating the above problems from both model architecture and data augmentation perspectives. Concretely, we build our model upon the advanced Transformer model. In order to enhance the discrimination of queries, a variety of external features (e.g., character, word, as well as script) are fed into the model and fused by a multi-scale attention mechanism. Moreover, to remedy the low resource challenge in this task, a novel machine translation–based strategy is proposed to automatically generate synthetic query-style data for low-resource languages. We contribute the first Q-LID test set called QID-21, which consists of search queries in 21 languages. Experimental results reveal that our model yields better classification accuracy than strong baselines and existing LID systems on both query and traditional LID tasks.1","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47340984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Lifelong Language Learning by Improving Pseudo-Sample Generation 通过改进伪样本生成,促进终身语言学习
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-06-30 DOI: 10.1162/coli_a_00449
Kasidis Kanwatchara, Thanapapas Horsuwan, Piyawat Lertvittayakumjorn, B. Kijsirikul, P. Vateekul
Abstract To achieve lifelong language learning, pseudo-rehearsal methods leverage samples generated from a language model to refresh the knowledge of previously learned tasks. Without proper controls, however, these methods could fail to retain the knowledge of complex tasks with longer texts since most of the generated samples are low in quality. To overcome the problem, we propose three specific contributions. First, we utilize double language models, each of which specializes in a specific part of the input, to produce high-quality pseudo samples. Second, we reduce the number of parameters used by applying adapter modules to enhance training efficiency. Third, we further improve the overall quality of pseudo samples using temporal ensembling and sample regeneration. The results show that our framework achieves significant improvement over baselines on multiple task sequences. Also, our pseudo sample analysis reveals helpful insights for designing even better pseudo-rehearsal methods in the future.
摘要为了实现终身语言学习,伪排练方法利用语言模型生成的样本来刷新先前学习任务的知识。然而,如果没有适当的控制,这些方法可能无法保留具有较长文本的复杂任务的知识,因为大多数生成的样本质量较低。为了克服这个问题,我们提出了三项具体贡献。首先,我们使用双语言模型,每个模型专门处理输入的特定部分,以生成高质量的伪样本。其次,我们通过应用适配器模块来减少使用的参数数量,以提高训练效率。第三,我们使用时间集成和样本再生进一步提高了伪样本的整体质量。结果表明,我们的框架在多个任务序列上实现了比基线的显著改进。此外,我们的伪样本分析为未来设计更好的伪排练方法提供了有益的见解。
{"title":"Enhancing Lifelong Language Learning by Improving Pseudo-Sample Generation","authors":"Kasidis Kanwatchara, Thanapapas Horsuwan, Piyawat Lertvittayakumjorn, B. Kijsirikul, P. Vateekul","doi":"10.1162/coli_a_00449","DOIUrl":"https://doi.org/10.1162/coli_a_00449","url":null,"abstract":"Abstract To achieve lifelong language learning, pseudo-rehearsal methods leverage samples generated from a language model to refresh the knowledge of previously learned tasks. Without proper controls, however, these methods could fail to retain the knowledge of complex tasks with longer texts since most of the generated samples are low in quality. To overcome the problem, we propose three specific contributions. First, we utilize double language models, each of which specializes in a specific part of the input, to produce high-quality pseudo samples. Second, we reduce the number of parameters used by applying adapter modules to enhance training efficiency. Third, we further improve the overall quality of pseudo samples using temporal ensembling and sample regeneration. The results show that our framework achieves significant improvement over baselines on multiple task sequences. Also, our pseudo sample analysis reveals helpful insights for designing even better pseudo-rehearsal methods in the future.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46164760","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Dimensional Modeling of Emotions in Text with Appraisal Theories: Corpus Creation, Annotation Reliability, and Prediction 基于评价理论的文本情感维度建模:语料库创建、标注可靠性与预测
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-06-10 DOI: 10.1162/coli_a_00461
Enrica Troiano, Laura Oberländer, Roman Klinger
The most prominent tasks in emotion analysis are to assign emotions to texts and to understand how emotions manifest in language. An important observation for natural language processing is that emotions can be communicated implicitly by referring to events alone, appealing to an empathetic, intersubjective understanding of events, even without explicitly mentioning an emotion name. In psychology, the class of emotion theories known as appraisal theories aims at explaining the link between events and emotions. Appraisals can be formalized as variables that measure a cognitive evaluation by people living through an event that they consider relevant. They include the assessment if an event is novel, if the person considers themselves to be responsible, if it is in line with their own goals, and so forth. Such appraisals explain which emotions are developed based on an event, for example, that a novel situation can induce surprise or one with uncertain consequences could evoke fear. We analyze the suitability of appraisal theories for emotion analysis in text with the goal of understanding if appraisal concepts can reliably be reconstructed by annotators, if they can be predicted by text classifiers, and if appraisal concepts help to identify emotion categories. To achieve that, we compile a corpus by asking people to textually describe events that triggered particular emotions and to disclose their appraisals. Then, we ask readers to reconstruct emotions and appraisals from the text. This set-up allows us to measure if emotions and appraisals can be recovered purely from text and provides a human baseline to judge a model’s performance measures. Our comparison of text classification methods to human annotators shows that both can reliably detect emotions and appraisals with similar performance. Therefore, appraisals constitute an alternative computational emotion analysis paradigm and further improve the categorization of emotions in text with joint models.
情绪分析中最突出的任务是将情绪分配给文本,并理解情绪如何在语言中表现出来。自然语言处理的一个重要观察结果是,情绪可以通过单独提及事件来隐含地传达,从而吸引对事件的移情、主体间理解,即使没有明确提及情绪名称。在心理学中,一类被称为评价理论的情绪理论旨在解释事件和情绪之间的联系。评估可以被形式化为变量,衡量人们在他们认为相关的事件中的认知评估。它们包括评估事件是否新颖,个人是否认为自己有责任,是否符合自己的目标,等等。这种评价解释了哪些情绪是基于一个事件产生的,例如,一个新颖的情况可能会引起惊讶,或者一个后果不确定的情况可能引起恐惧。我们分析了评价理论在文本中用于情绪分析的适用性,目的是了解评价概念是否可以由注释者可靠地重构,是否可以由文本分类器预测,以及评价概念是否有助于识别情绪类别。为了实现这一点,我们编写了一个语料库,要求人们用文本描述引发特定情绪的事件,并披露他们的评价。然后,我们要求读者从文本中重建情感和评价。这种设置使我们能够衡量情绪和评价是否可以纯粹从文本中恢复,并为判断模型的性能指标提供了一个人类基线。我们将文本分类方法与人类注释器进行了比较,结果表明两者都可以可靠地检测出具有相似性能的情绪和评价。因此,评价构成了一种替代的计算情感分析范式,并通过联合模型进一步改进了文本中情感的分类。
{"title":"Dimensional Modeling of Emotions in Text with Appraisal Theories: Corpus Creation, Annotation Reliability, and Prediction","authors":"Enrica Troiano, Laura Oberländer, Roman Klinger","doi":"10.1162/coli_a_00461","DOIUrl":"https://doi.org/10.1162/coli_a_00461","url":null,"abstract":"The most prominent tasks in emotion analysis are to assign emotions to texts and to understand how emotions manifest in language. An important observation for natural language processing is that emotions can be communicated implicitly by referring to events alone, appealing to an empathetic, intersubjective understanding of events, even without explicitly mentioning an emotion name. In psychology, the class of emotion theories known as appraisal theories aims at explaining the link between events and emotions. Appraisals can be formalized as variables that measure a cognitive evaluation by people living through an event that they consider relevant. They include the assessment if an event is novel, if the person considers themselves to be responsible, if it is in line with their own goals, and so forth. Such appraisals explain which emotions are developed based on an event, for example, that a novel situation can induce surprise or one with uncertain consequences could evoke fear. We analyze the suitability of appraisal theories for emotion analysis in text with the goal of understanding if appraisal concepts can reliably be reconstructed by annotators, if they can be predicted by text classifiers, and if appraisal concepts help to identify emotion categories. To achieve that, we compile a corpus by asking people to textually describe events that triggered particular emotions and to disclose their appraisals. Then, we ask readers to reconstruct emotions and appraisals from the text. This set-up allows us to measure if emotions and appraisals can be recovered purely from text and provides a human baseline to judge a model’s performance measures. Our comparison of text classification methods to human annotators shows that both can reliably detect emotions and appraisals with similar performance. Therefore, appraisals constitute an alternative computational emotion analysis paradigm and further improve the categorization of emotions in text with joint models.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43777842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
期刊
Computational Linguistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1