首页 > 最新文献

Computational Linguistics最新文献

英文 中文
Annotation Curricula to Implicitly Train Non-Expert Annotators 隐式训练非专家注释者的注释课程
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-06-04 DOI: 10.1162/coli_a_00436
Ji-Ung Lee, Jan-Christoph Klie, Iryna Gurevych
Abstract Annotation studies often require annotators to familiarize themselves with the task, its annotation scheme, and the data domain. This can be overwhelming in the beginning, mentally taxing, and induce errors into the resulting annotations; especially in citizen science or crowdsourcing scenarios where domain expertise is not required. To alleviate these issues, this work proposes annotation curricula, a novel approach to implicitly train annotators. The goal is to gradually introduce annotators into the task by ordering instances to be annotated according to a learning curriculum. To do so, this work formalizes annotation curricula for sentence- and paragraph-level annotation tasks, defines an ordering strategy, and identifies well-performing heuristics and interactively trained models on three existing English datasets. Finally, we provide a proof of concept for annotation curricula in a carefully designed user study with 40 voluntary participants who are asked to identify the most fitting misconception for English tweets about the Covid-19 pandemic. The results indicate that using a simple heuristic to order instances can already significantly reduce the total annotation time while preserving a high annotation quality. Annotation curricula thus can be a promising research direction to improve data collection. To facilitate future research—for instance, to adapt annotation curricula to specific tasks and expert annotation scenarios—all code and data from the user study consisting of 2,400 annotations is made available.1
摘要注释研究通常要求注释者熟悉任务、注释方案和数据域。这在一开始可能会让人不知所措,耗费大量精力,并导致注释出错;尤其是在不需要领域专业知识的公民科学或众包场景中。为了缓解这些问题,这项工作提出了注释课程,这是一种隐含训练注释者的新方法。目标是通过根据学习课程对要注释的实例进行排序,逐步将注释器引入到任务中。为此,这项工作正式化了句子和段落级注释任务的注释课程,定义了排序策略,并在三个现有的英语数据集上确定了性能良好的启发式方法和交互式训练的模型。最后,我们在一项精心设计的用户研究中为注释课程提供了概念验证,该研究有40名自愿参与者,他们被要求确定关于新冠肺炎大流行的英语推文最合适的误解。结果表明,使用简单的启发式排序实例已经可以显著减少总的注释时间,同时保持高的注释质量。因此,注释课程可以成为改进数据收集的一个很有前途的研究方向。为了促进未来的研究——例如,使注释课程适应特定任务和专家注释场景——由2400个注释组成的用户研究的所有代码和数据都可用。1
{"title":"Annotation Curricula to Implicitly Train Non-Expert Annotators","authors":"Ji-Ung Lee, Jan-Christoph Klie, Iryna Gurevych","doi":"10.1162/coli_a_00436","DOIUrl":"https://doi.org/10.1162/coli_a_00436","url":null,"abstract":"Abstract Annotation studies often require annotators to familiarize themselves with the task, its annotation scheme, and the data domain. This can be overwhelming in the beginning, mentally taxing, and induce errors into the resulting annotations; especially in citizen science or crowdsourcing scenarios where domain expertise is not required. To alleviate these issues, this work proposes annotation curricula, a novel approach to implicitly train annotators. The goal is to gradually introduce annotators into the task by ordering instances to be annotated according to a learning curriculum. To do so, this work formalizes annotation curricula for sentence- and paragraph-level annotation tasks, defines an ordering strategy, and identifies well-performing heuristics and interactively trained models on three existing English datasets. Finally, we provide a proof of concept for annotation curricula in a carefully designed user study with 40 voluntary participants who are asked to identify the most fitting misconception for English tweets about the Covid-19 pandemic. The results indicate that using a simple heuristic to order instances can already significantly reduce the total annotation time while preserving a high annotation quality. Annotation curricula thus can be a promising research direction to improve data collection. To facilitate future research—for instance, to adapt annotation curricula to specific tasks and expert annotation scenarios—all code and data from the user study consisting of 2,400 annotations is made available.1","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44039633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Universal Discourse Representation Structure Parsing 通用语篇表示结构分析
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-05-20 DOI: 10.1162/coli_a_00406
Jiangming Liu, Shay B. Cohen, Mirella Lapata, Johan Bos
Abstract We consider the task of crosslingual semantic parsing in the style of Discourse Representation Theory (DRT) where knowledge from annotated corpora in a resource-rich language is transferred via bitext to guide learning in other languages. We introduce 𝕌niversal Discourse Representation Theory (𝕌DRT), a variant of DRT that explicitly anchors semantic representations to tokens in the linguistic input. We develop a semantic parsing framework based on the Transformer architecture and utilize it to obtain semantic resources in multiple languages following two learning schemes. The many-to-one approach translates non-English text to English, and then runs a relatively accurate English parser on the translated text, while the one-to-many approach translates gold standard English to non-English text and trains multiple parsers (one per language) on the translations. Experimental results on the Parallel Meaning Bank show that our proposal outperforms strong baselines by a wide margin and can be used to construct (silver-standard) meaning banks for 99 languages.
摘要本文研究了基于话语表示理论(Discourse Representation Theory, DRT)的跨语言语义分析任务,即资源丰富的语言中带注释的语料库中的知识通过文本传递,以指导其他语言的学习。我们介绍𝕌niversal话语表示理论(𝕌DRT),这是DRT的一种变体,它显式地将语义表示锚定在语言输入中的标记上。我们开发了一个基于Transformer架构的语义解析框架,并利用它在两种学习方案下获取多语言的语义资源。多对一方法将非英语文本翻译成英语,然后在翻译的文本上运行一个相对准确的英语解析器,而一对多方法将黄金标准英语翻译成非英语文本,并在翻译上训练多个解析器(每种语言一个)。在平行意义库上的实验结果表明,我们的提议在很大程度上优于强基线,并且可以用于构建99种语言的(银标准)意义库。
{"title":"Universal Discourse Representation Structure Parsing","authors":"Jiangming Liu, Shay B. Cohen, Mirella Lapata, Johan Bos","doi":"10.1162/coli_a_00406","DOIUrl":"https://doi.org/10.1162/coli_a_00406","url":null,"abstract":"Abstract We consider the task of crosslingual semantic parsing in the style of Discourse Representation Theory (DRT) where knowledge from annotated corpora in a resource-rich language is transferred via bitext to guide learning in other languages. We introduce 𝕌niversal Discourse Representation Theory (𝕌DRT), a variant of DRT that explicitly anchors semantic representations to tokens in the linguistic input. We develop a semantic parsing framework based on the Transformer architecture and utilize it to obtain semantic resources in multiple languages following two learning schemes. The many-to-one approach translates non-English text to English, and then runs a relatively accurate English parser on the translated text, while the one-to-many approach translates gold standard English to non-English text and trains multiple parsers (one per language) on the translations. Experimental results on the Parallel Meaning Bank show that our proposal outperforms strong baselines by a wide margin and can be used to construct (silver-standard) meaning banks for 99 languages.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48946353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Certified Robustness to Text Adversarial Attacks by Randomized [MASK] 基于随机[MASK]的文本对抗性攻击认证鲁棒性
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-05-08 DOI: 10.1162/coli_a_00476
Jiehang Zeng, Xiaoqing Zheng, Jianhan Xu, Linyang Li, Liping Yuan, Xuanjing Huang
Very recently, few certified defense methods have been developed to provably guarantee the robustness of a text classifier to adversarial synonym substitutions. However, all the existing certified defense methods assume that the defenders have been informed of how the adversaries generate synonyms, which is not a realistic scenario. In this study, we propose a certifiably robust defense method by randomly masking a certain proportion of the words in an input text, in which the above unrealistic assumption is no longer necessary. The proposed method can defend against not only word substitution-based attacks, but also character-level perturbations. We can certify the classifications of over 50% of texts to be robust to any perturbation of five words on AGNEWS, and two words on SST2 dataset. The experimental results show that our randomized smoothing method significantly outperforms recently proposed defense methods across multiple datasets under different attack algorithms.
最近,很少有经过认证的防御方法被开发出来,以可证明地保证文本分类器对对抗性同义词替换的鲁棒性。然而,所有现有的认证防御方法都假设防御者已经被告知对手如何生成同义词,这不是一个现实的场景。在这项研究中,我们提出了一种可证明稳健的防御方法,通过随机屏蔽输入文本中一定比例的单词,其中不再需要上述不切实际的假设。所提出的方法不仅可以抵御基于单词替换的攻击,还可以抵御字符级的扰动。我们可以证明超过50%的文本的分类对AGNEWS上的五个单词和SST2数据集上的两个单词的任何扰动都是稳健的。实验结果表明,在不同的攻击算法下,我们的随机平滑方法在多个数据集上显著优于最近提出的防御方法。
{"title":"Certified Robustness to Text Adversarial Attacks by Randomized [MASK]","authors":"Jiehang Zeng, Xiaoqing Zheng, Jianhan Xu, Linyang Li, Liping Yuan, Xuanjing Huang","doi":"10.1162/coli_a_00476","DOIUrl":"https://doi.org/10.1162/coli_a_00476","url":null,"abstract":"Very recently, few certified defense methods have been developed to provably guarantee the robustness of a text classifier to adversarial synonym substitutions. However, all the existing certified defense methods assume that the defenders have been informed of how the adversaries generate synonyms, which is not a realistic scenario. In this study, we propose a certifiably robust defense method by randomly masking a certain proportion of the words in an input text, in which the above unrealistic assumption is no longer necessary. The proposed method can defend against not only word substitution-based attacks, but also character-level perturbations. We can certify the classifications of over 50% of texts to be robust to any perturbation of five words on AGNEWS, and two words on SST2 dataset. The experimental results show that our randomized smoothing method significantly outperforms recently proposed defense methods across multiple datasets under different attack algorithms.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44285495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Kathy McKeown Interviews Bonnie Webber Kathy McKeown采访Bonnie Webber
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-04-01 DOI: 10.1162/coli_a_00393
B. Webber
Abstract Because the 2020 ACL Lifetime Achievement Award presentation could not be done in person, we replaced the usual LTA talk with an interview between Professor Kathy McKeown (Columbia University) and the recipient, Bonnie Webber. The following is an edited version of the interview, with added citations.
摘要由于2020年ACL终身成就奖颁奖典礼无法亲自进行,我们用哥伦比亚大学Kathy McKeown教授和获奖者Bonnie Webber之间的采访取代了通常的LTA演讲。以下是采访的编辑版本,并添加了引文。
{"title":"Kathy McKeown Interviews Bonnie Webber","authors":"B. Webber","doi":"10.1162/coli_a_00393","DOIUrl":"https://doi.org/10.1162/coli_a_00393","url":null,"abstract":"Abstract Because the 2020 ACL Lifetime Achievement Award presentation could not be done in person, we replaced the usual LTA talk with an interview between Professor Kathy McKeown (Columbia University) and the recipient, Bonnie Webber. The following is an edited version of the interview, with added citations.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48392240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Depth-Bounded Statistical PCFG Induction as a Model of Human Grammar Acquisition 深度有界统计PCFG归纳法作为人类语法习得模型
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-04-01 DOI: 10.1162/coli_a_00399
Lifeng Jin, Lane Schwartz, F. Doshi-Velez, Timothy A. Miller, William Schuler
Abstract This article describes a simple PCFG induction model with a fixed category domain that predicts a large majority of attested constituent boundaries, and predicts labels consistent with nearly half of attested constituent labels on a standard evaluation data set of child-directed speech. The article then explores the idea that the difference between simple grammars exhibited by child learners and fully recursive grammars exhibited by adult learners may be an effect of increasing working memory capacity, where the shallow grammars are constrained images of the recursive grammars. An implementation of these memory bounds as limits on center embedding in a depth-specific transform of a recursive grammar yields a significant improvement over an equivalent but unbounded baseline, suggesting that this arrangement may indeed confer a learning advantage.
本文描述了一个简单的PCFG归纳模型,该模型具有固定的类别域,可以预测绝大多数已证实的成分边界,并在儿童导向语音的标准评价数据集上预测出与近一半已证实的成分标签一致的标签。然后,文章探讨了儿童学习者表现出的简单语法和成人学习者表现出的完全递归语法之间的差异可能是工作记忆容量增加的影响,其中浅层语法是递归语法的约束图像。在递归语法的深度特定转换中,将这些内存边界作为中心嵌入限制的实现,会比等效的无界基线产生显著的改进,这表明这种安排可能确实具有学习优势。
{"title":"Depth-Bounded Statistical PCFG Induction as a Model of Human Grammar Acquisition","authors":"Lifeng Jin, Lane Schwartz, F. Doshi-Velez, Timothy A. Miller, William Schuler","doi":"10.1162/coli_a_00399","DOIUrl":"https://doi.org/10.1162/coli_a_00399","url":null,"abstract":"Abstract This article describes a simple PCFG induction model with a fixed category domain that predicts a large majority of attested constituent boundaries, and predicts labels consistent with nearly half of attested constituent labels on a standard evaluation data set of child-directed speech. The article then explores the idea that the difference between simple grammars exhibited by child learners and fully recursive grammars exhibited by adult learners may be an effect of increasing working memory capacity, where the shallow grammars are constrained images of the recursive grammars. An implementation of these memory bounds as limits on center embedding in a depth-specific transform of a recursive grammar yields a significant improvement over an equivalent but unbounded baseline, suggesting that this arrangement may indeed confer a learning advantage.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45508651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Formal Basis of a Language Universal 通用语言的形式基础
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-04-01 DOI: 10.1162/coli_a_00394
M. Stanojevic, M. Steedman
Abstract Steedman (2020) proposes as a formal universal of natural language grammar that grammatical permutations of the kind that have given rise to transformational rules are limited to a class known to mathematicians and computer scientists as the “separable” permutations. This class of permutations is exactly the class that can be expressed in combinatory categorial grammars (CCGs). The excluded non-separable permutations do in fact seem to be absent in a number of studies of crosslinguistic variation in word order in nominal and verbal constructions. The number of permutations that are separable grows in the number n of lexical elements in the construction as the Large Schröder Number Sn−1. Because that number grows much more slowly than the n! number of all permutations, this generalization is also of considerable practical interest for computational applications such as parsing and machine translation. The present article examines the mathematical and computational origins of this restriction, and the reason it is exactly captured in CCG without the imposition of any further constraints.
Steedman(2020)提出,作为自然语言语法的一种形式普遍性,产生转换规则的那种语法排列仅限于数学家和计算机科学家称为“可分离”排列的一类。这类排列正是可以用组合范畴语法(ccg)表示的一类。被排除在外的不可分排列实际上在许多关于词序在名义和言语结构中的跨语言变化的研究中似乎并不存在。在结构中,可分离排列的数目随着词法元素的数目n的增加而增加,如Schröder number Sn−1。因为这个数的增长速度比n慢得多!所有排列的数目,这种泛化对于解析和机器翻译等计算应用也有相当大的实际意义。本文考察了这一限制的数学和计算起源,以及在没有施加任何进一步约束的情况下在CCG中准确捕获它的原因。
{"title":"Formal Basis of a Language Universal","authors":"M. Stanojevic, M. Steedman","doi":"10.1162/coli_a_00394","DOIUrl":"https://doi.org/10.1162/coli_a_00394","url":null,"abstract":"Abstract Steedman (2020) proposes as a formal universal of natural language grammar that grammatical permutations of the kind that have given rise to transformational rules are limited to a class known to mathematicians and computer scientists as the “separable” permutations. This class of permutations is exactly the class that can be expressed in combinatory categorial grammars (CCGs). The excluded non-separable permutations do in fact seem to be absent in a number of studies of crosslinguistic variation in word order in nominal and verbal constructions. The number of permutations that are separable grows in the number n of lexical elements in the construction as the Large Schröder Number Sn−1. Because that number grows much more slowly than the n! number of all permutations, this generalization is also of considerable practical interest for computational applications such as parsing and machine translation. The present article examines the mathematical and computational origins of this restriction, and the reason it is exactly captured in CCG without the imposition of any further constraints.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42863124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Comparing Knowledge-Intensive and Data-Intensive Models for English Resource Semantic Parsing 英语资源语义分析的知识密集型和数据密集型模型比较
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-04-01 DOI: 10.1162/coli_a_00395
Junjie Cao, Zi-yu Lin, Weiwei Sun, Xiaojun Wan
Abstract In this work, we present a phenomenon-oriented comparative analysis of the two dominant approaches in English Resource Semantic (ERS) parsing: classic, knowledge-intensive and neural, data-intensive models. To reflect state-of-the-art neural NLP technologies, a factorization-based parser is introduced that can produce Elementary Dependency Structures much more accurately than previous data-driven parsers. We conduct a suite of tests for different linguistic phenomena to analyze the grammatical competence of different parsers, where we show that, despite comparable performance overall, knowledge- and data-intensive models produce different types of errors, in a way that can be explained by their theoretical properties. This analysis is beneficial to in-depth evaluation of several representative parsing techniques and leads to new directions for parser development.
摘要在这项工作中,我们对英语资源语义分析中的两种主要方法进行了面向现象的比较分析:经典的知识密集型和神经的数据密集型模型。为了反映最先进的神经NLP技术,引入了一种基于因子分解的解析器,该解析器可以比以前的数据驱动解析器更准确地生成基本依赖结构。我们对不同的语言现象进行了一系列测试,以分析不同语法分析器的语法能力,我们发现,尽管总体性能相当,但知识和数据密集型模型会产生不同类型的错误,这可以用它们的理论特性来解释。该分析有利于深入评估几种具有代表性的解析技术,并为解析器的开发开辟了新的方向。
{"title":"Comparing Knowledge-Intensive and Data-Intensive Models for English Resource Semantic Parsing","authors":"Junjie Cao, Zi-yu Lin, Weiwei Sun, Xiaojun Wan","doi":"10.1162/coli_a_00395","DOIUrl":"https://doi.org/10.1162/coli_a_00395","url":null,"abstract":"Abstract In this work, we present a phenomenon-oriented comparative analysis of the two dominant approaches in English Resource Semantic (ERS) parsing: classic, knowledge-intensive and neural, data-intensive models. To reflect state-of-the-art neural NLP technologies, a factorization-based parser is introduced that can produce Elementary Dependency Structures much more accurately than previous data-driven parsers. We conduct a suite of tests for different linguistic phenomena to analyze the grammatical competence of different parsers, where we show that, despite comparable performance overall, knowledge- and data-intensive models produce different types of errors, in a way that can be explained by their theoretical properties. This analysis is beneficial to in-depth evaluation of several representative parsing techniques and leads to new directions for parser development.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47809438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Python for Linguists 面向语言学家的Python
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-04-01 DOI: 10.1162/coli_r_00400
Benjamin Roth, Michael Wiegand
Teaching programming skills is a hard task. It is even harder if one targets an audience with no or little mathematical background. Although there are books on programming that target such groups, they often fail to raise or maintain interest due to artificial examples that lack reference to the professional issues that the audience typically face. This book fills the gap by addressing linguistics, a profession and academic subject for which basic knowledge of script programming is becoming more and more important. The book Python for Linguists by Michael Hammond is an introductory Python course targeted at linguists with no prior programming background. It succeeds previous books for Perl (Hammond 2008) and Java (Hammond 2002) by the same author, and reflects the current de facto prevalence of Python when it comes to adoption and available packages for natural language processing. We feel it necessary to clarify that the book aims at (general) linguists in the broad sense rather than computational linguists. Its aim is to teach linguists the fundamental concepts of programming using typical examples from linguistics. The book should not be mistaken as a course for learning basic algorithms in computational linguistics. We acknowledge that the author nowhere makes such a claim; however, given the thematic proximity to computational linguistics, one should have the right expectation before working with the book. Chapters 1–5 lay the foundations of the Python programming language, introducing the most important language constructs but deferring object oriented programming to a later part of the book. The focus in Chapters 1 and 2 covers the basic data types (numbers, strings, dictionaries), with a particular emphasis on simple string operations, and introduces some more advanced concepts such as mutability. Chapters 3–5 introduce control structures, input–output operations, and modules. The book goes at great length to visualize the program flow and the state of different variables for different steps in a program execution, which is certainly very helpful for learners with no prior programming experience. The book also guides the learner to understand certain error types that frequently occur in computer programming (but might be unintuitive for beginners). For example, when discussing function calls, much care is devoted to pointing out the unintended consequences stemming from mutability and side effects.
教授编程技能是一项艰巨的任务。如果目标受众没有或几乎没有数学背景,那就更难了。尽管有一些针对这类群体的编程书籍,但由于缺乏提及观众通常面临的专业问题的人为例子,它们往往无法引起或保持兴趣。这本书填补了语言学的空白,语言学是一门专业和学术学科,脚本编程的基础知识对它来说越来越重要。迈克尔·哈蒙德的《语言学家的Python》一书是一门面向没有编程背景的语言学家的Python入门课程。它继承了同一作者以前的Perl(Hammond 2008)和Java(Hammond 2002)的著作,并反映了Python在采用和可用于自然语言处理的包方面的实际流行情况。我们认为有必要澄清,这本书针对的是广义的(一般)语言学家,而不是计算语言学家。其目的是用语言学中的典型例子向语言学家传授编程的基本概念。这本书不应该被误认为是一门学习计算语言学基本算法的课程。我们承认,提交人没有提出这样的主张;然而,考虑到主题接近计算语言学,在阅读这本书之前应该有正确的期望。第1-5章奠定了Python编程语言的基础,介绍了最重要的语言结构,但将面向对象编程推迟到本书的后面部分。第1章和第2章的重点涵盖了基本的数据类型(数字、字符串、字典),特别强调简单的字符串操作,并引入了一些更高级的概念,如可变性。第3-5章介绍了控制结构、输入输出操作和模块。这本书花了很大的篇幅来可视化程序流程和程序执行中不同步骤的不同变量的状态,这对没有编程经验的学习者肯定非常有帮助。这本书还指导学习者理解计算机编程中经常出现的某些错误类型(但对初学者来说可能是不直观的)。例如,在讨论函数调用时,要特别注意指出由可变性和副作用引起的意外后果。
{"title":"Python for Linguists","authors":"Benjamin Roth, Michael Wiegand","doi":"10.1162/coli_r_00400","DOIUrl":"https://doi.org/10.1162/coli_r_00400","url":null,"abstract":"Teaching programming skills is a hard task. It is even harder if one targets an audience with no or little mathematical background. Although there are books on programming that target such groups, they often fail to raise or maintain interest due to artificial examples that lack reference to the professional issues that the audience typically face. This book fills the gap by addressing linguistics, a profession and academic subject for which basic knowledge of script programming is becoming more and more important. The book Python for Linguists by Michael Hammond is an introductory Python course targeted at linguists with no prior programming background. It succeeds previous books for Perl (Hammond 2008) and Java (Hammond 2002) by the same author, and reflects the current de facto prevalence of Python when it comes to adoption and available packages for natural language processing. We feel it necessary to clarify that the book aims at (general) linguists in the broad sense rather than computational linguists. Its aim is to teach linguists the fundamental concepts of programming using typical examples from linguistics. The book should not be mistaken as a course for learning basic algorithms in computational linguistics. We acknowledge that the author nowhere makes such a claim; however, given the thematic proximity to computational linguistics, one should have the right expectation before working with the book. Chapters 1–5 lay the foundations of the Python programming language, introducing the most important language constructs but deferring object oriented programming to a later part of the book. The focus in Chapters 1 and 2 covers the basic data types (numbers, strings, dictionaries), with a particular emphasis on simple string operations, and introduces some more advanced concepts such as mutability. Chapters 3–5 introduce control structures, input–output operations, and modules. The book goes at great length to visualize the program flow and the state of different variables for different steps in a program execution, which is certainly very helpful for learners with no prior programming experience. The book also guides the learner to understand certain error types that frequently occur in computer programming (but might be unintuitive for beginners). For example, when discussing function calls, much care is devoted to pointing out the unintended consequences stemming from mutability and side effects.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47640448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semantic Data Set Construction from Human Clustering and Spatial Arrangement 基于人聚类和空间排列的语义数据集构建
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-04-01 DOI: 10.1162/coli_a_00396
Olga Majewska, Diana McCarthy, Jasper J. F. van den Bosch, N. Kriegeskorte, Ivan Vulic, A. Korhonen
Abstract Research into representation learning models of lexical semantics usually utilizes some form of intrinsic evaluation to ensure that the learned representations reflect human semantic judgments. Lexical semantic similarity estimation is a widely used evaluation method, but efforts have typically focused on pairwise judgments of words in isolation, or are limited to specific contexts and lexical stimuli. There are limitations with these approaches that either do not provide any context for judgments, and thereby ignore ambiguity, or provide very specific sentential contexts that cannot then be used to generate a larger lexical resource. Furthermore, similarity between more than two items is not considered. We provide a full description and analysis of our recently proposed methodology for large-scale data set construction that produces a semantic classification of a large sample of verbs in the first phase, as well as multi-way similarity judgments made within the resultant semantic classes in the second phase. The methodology uses a spatial multi-arrangement approach proposed in the field of cognitive neuroscience for capturing multi-way similarity judgments of visual stimuli. We have adapted this method to handle polysemous linguistic stimuli and much larger samples than previous work. We specifically target verbs, but the method can equally be applied to other parts of speech. We perform cluster analysis on the data from the first phase and demonstrate how this might be useful in the construction of a comprehensive verb resource. We also analyze the semantic information captured by the second phase and discuss the potential of the spatially induced similarity judgments to better reflect human notions of word similarity. We demonstrate how the resultant data set can be used for fine-grained analyses and evaluation of representation learning models on the intrinsic tasks of semantic clustering and semantic similarity. In particular, we find that stronger static word embedding methods still outperform lexical representations emerging from more recent pre-training methods, both on word-level similarity and clustering. Moreover, thanks to the data set’s vast coverage, we are able to compare the benefits of specializing vector representations for a particular type of external knowledge by evaluating FrameNet- and VerbNet-retrofitted models on specific semantic domains such as “Heat” or “Motion.”
摘要对词汇语义表征学习模型的研究通常利用某种形式的内在评价来确保所学习的表征反映人类的语义判断。词汇语义相似性估计是一种广泛使用的评估方法,但通常侧重于孤立地对单词进行成对判断,或者仅限于特定的上下文和词汇刺激。这些方法存在局限性,要么不提供任何上下文进行判断,从而忽略歧义,要么提供非常具体的句子上下文,然后无法用于生成更大的词汇资源。此外,不考虑两个以上项目之间的相似性。我们对我们最近提出的大规模数据集构建方法进行了全面的描述和分析,该方法在第一阶段产生了大量动词样本的语义分类,并在第二阶段对由此产生的语义类进行了多向相似性判断。该方法使用认知神经科学领域提出的空间多重排列方法来捕捉视觉刺激的多向相似性判断。我们已经将这种方法应用于处理多义词的语言刺激和比以前的工作大得多的样本。我们专门针对动词,但这种方法同样适用于其他词性。我们对第一阶段的数据进行了聚类分析,并展示了这在构建综合动词资源中是如何有用的。我们还分析了第二阶段捕获的语义信息,并讨论了空间诱导的相似性判断的潜力,以更好地反映人类对单词相似性的概念。我们展示了如何将生成的数据集用于对语义聚类和语义相似性的内在任务的表示学习模型进行细粒度分析和评估。特别是,我们发现更强的静态单词嵌入方法在单词级别的相似性和聚类方面仍然优于最近的预训练方法中出现的词汇表示。此外,由于数据集的广泛覆盖,我们能够通过评估特定语义域(如“热”或“运动”)上的FrameNet和VerbNet改进模型,来比较专门化特定类型外部知识的向量表示的好处
{"title":"Semantic Data Set Construction from Human Clustering and Spatial Arrangement","authors":"Olga Majewska, Diana McCarthy, Jasper J. F. van den Bosch, N. Kriegeskorte, Ivan Vulic, A. Korhonen","doi":"10.1162/coli_a_00396","DOIUrl":"https://doi.org/10.1162/coli_a_00396","url":null,"abstract":"Abstract Research into representation learning models of lexical semantics usually utilizes some form of intrinsic evaluation to ensure that the learned representations reflect human semantic judgments. Lexical semantic similarity estimation is a widely used evaluation method, but efforts have typically focused on pairwise judgments of words in isolation, or are limited to specific contexts and lexical stimuli. There are limitations with these approaches that either do not provide any context for judgments, and thereby ignore ambiguity, or provide very specific sentential contexts that cannot then be used to generate a larger lexical resource. Furthermore, similarity between more than two items is not considered. We provide a full description and analysis of our recently proposed methodology for large-scale data set construction that produces a semantic classification of a large sample of verbs in the first phase, as well as multi-way similarity judgments made within the resultant semantic classes in the second phase. The methodology uses a spatial multi-arrangement approach proposed in the field of cognitive neuroscience for capturing multi-way similarity judgments of visual stimuli. We have adapted this method to handle polysemous linguistic stimuli and much larger samples than previous work. We specifically target verbs, but the method can equally be applied to other parts of speech. We perform cluster analysis on the data from the first phase and demonstrate how this might be useful in the construction of a comprehensive verb resource. We also analyze the semantic information captured by the second phase and discuss the potential of the spatially induced similarity judgments to better reflect human notions of word similarity. We demonstrate how the resultant data set can be used for fine-grained analyses and evaluation of representation learning models on the intrinsic tasks of semantic clustering and semantic similarity. In particular, we find that stronger static word embedding methods still outperform lexical representations emerging from more recent pre-training methods, both on word-level similarity and clustering. Moreover, thanks to the data set’s vast coverage, we are able to compare the benefits of specializing vector representations for a particular type of external knowledge by evaluating FrameNet- and VerbNet-retrofitted models on specific semantic domains such as “Heat” or “Motion.”","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48554442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Analysis and Evaluation of Language Models for Word Sense Disambiguation 词义消歧的语言模型分析与评价
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2021-03-26 DOI: 10.1162/coli_a_00405
Daniel Loureiro, Kiamehr Rezaee, Mohammad Taher Pilehvar, José Camacho-Collados
Abstract Transformer-based language models have taken many fields in NLP by storm. BERT and its derivatives dominate most of the existing evaluation benchmarks, including those for Word Sense Disambiguation (WSD), thanks to their ability in capturing context-sensitive semantic nuances. However, there is still little knowledge about their capabilities and potential limitations in encoding and recovering word senses. In this article, we provide an in-depth quantitative and qualitative analysis of the celebrated BERT model with respect to lexical ambiguity. One of the main conclusions of our analysis is that BERT can accurately capture high-level sense distinctions, even when a limited number of examples is available for each word sense. Our analysis also reveals that in some cases language models come close to solving coarse-grained noun disambiguation under ideal conditions in terms of availability of training data and computing resources. However, this scenario rarely occurs in real-world settings and, hence, many practical challenges remain even in the coarse-grained setting. We also perform an in-depth comparison of the two main language model-based WSD strategies, namely, fine-tuning and feature extraction, finding that the latter approach is more robust with respect to sense bias and it can better exploit limited available training data. In fact, the simple feature extraction strategy of averaging contextualized embeddings proves robust even using only three training sentences per word sense, with minimal improvements obtained by increasing the size of this training data.
基于变换的语言模型在自然语言处理领域掀起了一股热潮。BERT及其衍生工具由于能够捕获上下文敏感的语义细微差别,在大多数现有的评估基准中占据主导地位,包括那些用于词义消歧(WSD)的基准。然而,关于它们在编码和恢复词义方面的能力和潜在局限性,我们仍然知之甚少。在本文中,我们对著名的BERT模型在词汇歧义方面进行了深入的定量和定性分析。我们分析的一个主要结论是,BERT可以准确地捕获高级意义的区别,即使每个词义的可用示例数量有限。我们的分析还表明,在某些情况下,就训练数据和计算资源的可用性而言,语言模型在理想条件下接近于解决粗粒度名词消歧问题。然而,这种情况很少出现在现实环境中,因此,即使在粗粒度的环境中也存在许多实际挑战。我们还对两种主要的基于语言模型的WSD策略(即微调和特征提取)进行了深入的比较,发现后一种方法在感觉偏差方面更具鲁棒性,并且可以更好地利用有限的可用训练数据。事实上,即使每个词义只使用三个训练句子,平均上下文化嵌入的简单特征提取策略也证明了鲁棒性,并且通过增加训练数据的大小获得了最小的改进。
{"title":"Analysis and Evaluation of Language Models for Word Sense Disambiguation","authors":"Daniel Loureiro, Kiamehr Rezaee, Mohammad Taher Pilehvar, José Camacho-Collados","doi":"10.1162/coli_a_00405","DOIUrl":"https://doi.org/10.1162/coli_a_00405","url":null,"abstract":"Abstract Transformer-based language models have taken many fields in NLP by storm. BERT and its derivatives dominate most of the existing evaluation benchmarks, including those for Word Sense Disambiguation (WSD), thanks to their ability in capturing context-sensitive semantic nuances. However, there is still little knowledge about their capabilities and potential limitations in encoding and recovering word senses. In this article, we provide an in-depth quantitative and qualitative analysis of the celebrated BERT model with respect to lexical ambiguity. One of the main conclusions of our analysis is that BERT can accurately capture high-level sense distinctions, even when a limited number of examples is available for each word sense. Our analysis also reveals that in some cases language models come close to solving coarse-grained noun disambiguation under ideal conditions in terms of availability of training data and computing resources. However, this scenario rarely occurs in real-world settings and, hence, many practical challenges remain even in the coarse-grained setting. We also perform an in-depth comparison of the two main language model-based WSD strategies, namely, fine-tuning and feature extraction, finding that the latter approach is more robust with respect to sense bias and it can better exploit limited available training data. In fact, the simple feature extraction strategy of averaging contextualized embeddings proves robust even using only three training sentences per word sense, with minimal improvements obtained by increasing the size of this training data.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2021-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"64495119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
期刊
Computational Linguistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1