首页 > 最新文献

Computational Linguistics最新文献

英文 中文
Annotation Error Detection: Analyzing the Past and Present for a More Coherent Future 注释错误检测:分析过去和现在以获得更连贯的未来
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-06-05 DOI: 10.1162/coli_a_00464
Jan-Christoph Klie, B. Webber, Iryna Gurevych
Annotated data is an essential ingredient in natural language processing for training and evaluating machine learning models. It is therefore very desirable for the annotations to be of high quality. Recent work, however, has shown that several popular datasets contain a surprising number of annotation errors or inconsistencies. To alleviate this issue, many methods for annotation error detection have been devised over the years. While researchers show that their approaches work well on their newly introduced datasets, they rarely compare their methods to previous work or on the same datasets. This raises strong concerns on methods’ general performance and makes it difficult to assess their strengths and weaknesses. We therefore reimplement 18 methods for detecting potential annotation errors and evaluate them on 9 English datasets for text classification as well as token and span labeling. In addition, we define a uniform evaluation setup including a new formalization of the annotation error detection task, evaluation protocol, and general best practices. To facilitate future research and reproducibility, we release our datasets and implementations in an easy-to-use and open source software package.1
在训练和评估机器学习模型的自然语言处理中,带注释的数据是必不可少的组成部分。因此,高质量的注释是非常可取的。然而,最近的研究表明,一些流行的数据集包含了数量惊人的注释错误或不一致。为了缓解这个问题,多年来已经设计了许多注释错误检测方法。虽然研究人员表明他们的方法在新引入的数据集上工作得很好,但他们很少将他们的方法与以前的工作或相同的数据集进行比较。这引起了对方法一般性能的强烈关注,并使评估它们的优缺点变得困难。因此,我们重新实现了18种方法来检测潜在的注释错误,并在9个英语数据集上对它们进行评估,用于文本分类以及标记和span标记。此外,我们定义了一个统一的评估设置,包括注释错误检测任务的新形式化、评估协议和通用最佳实践。为了促进未来的研究和可重复性,我们在一个易于使用的开源软件包中发布了我们的数据集和实现
{"title":"Annotation Error Detection: Analyzing the Past and Present for a More Coherent Future","authors":"Jan-Christoph Klie, B. Webber, Iryna Gurevych","doi":"10.1162/coli_a_00464","DOIUrl":"https://doi.org/10.1162/coli_a_00464","url":null,"abstract":"Annotated data is an essential ingredient in natural language processing for training and evaluating machine learning models. It is therefore very desirable for the annotations to be of high quality. Recent work, however, has shown that several popular datasets contain a surprising number of annotation errors or inconsistencies. To alleviate this issue, many methods for annotation error detection have been devised over the years. While researchers show that their approaches work well on their newly introduced datasets, they rarely compare their methods to previous work or on the same datasets. This raises strong concerns on methods’ general performance and makes it difficult to assess their strengths and weaknesses. We therefore reimplement 18 methods for detecting potential annotation errors and evaluate them on 9 English datasets for text classification as well as token and span labeling. In addition, we define a uniform evaluation setup including a new formalization of the annotation error detection task, evaluation protocol, and general best practices. To facilitate future research and reproducibility, we release our datasets and implementations in an easy-to-use and open source software package.1","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42479264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Noun2Verb: Probabilistic Frame Semantics for Word Class Conversion 名词动词:类转换的概率框架语义
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-05-12 DOI: 10.1162/coli_a_00447
Lei Yu, Yang Xu
Abstract Humans can flexibly extend word usages across different grammatical classes, a phenomenon known as word class conversion. Noun-to-verb conversion, or denominal verb (e.g., to Google a cheap flight), is one of the most prevalent forms of word class conversion. However, existing natural language processing systems are impoverished in interpreting and generating novel denominal verb usages. Previous work has suggested that novel denominal verb usages are comprehensible if the listener can compute the intended meaning based on shared knowledge with the speaker. Here we explore a computational formalism for this proposal couched in frame semantics. We present a formal framework, Noun2Verb, that simulates the production and comprehension of novel denominal verb usages by modeling shared knowledge of speaker and listener in semantic frames. We evaluate an incremental set of probabilistic models that learn to interpret and generate novel denominal verb usages via paraphrasing. We show that a model where the speaker and listener cooperatively learn the joint distribution over semantic frame elements better explains the empirical denominal verb usages than state-of-the-art language models, evaluated against data from (1) contemporary English in both adult and child speech, (2) contemporary Mandarin Chinese, and (3) the historical development of English. Our work grounds word class conversion in probabilistic frame semantics and bridges the gap between natural language processing systems and humans in lexical creativity.
摘要人类可以灵活地将单词用法扩展到不同的语法类别,这种现象被称为单词类别转换。名词-动词转换,或名词-动词(例如,谷歌廉价航班),是最常见的词类转换形式之一。然而,现有的自然语言处理系统在解释和生成新的名词动词用法方面很差。先前的工作表明,如果听众能够根据与说话者共享的知识来计算意图,那么新的名词动词用法是可以理解的。在这里,我们探讨了这个建议的计算形式,用框架语义表达。我们提出了一个形式框架Noun2Verb,通过在语义框架中建模说话人和听话人的共享知识,模拟新名词动词用法的产生和理解。我们评估了一组递增的概率模型,这些模型学习通过转述来解释和生成新的名词动词用法。我们发现,说话者和听者合作学习语义框架元素的联合分布的模型比最先进的语言模型更好地解释了经验名词动词用法,该模型根据以下数据进行了评估:(1)成人和儿童言语中的当代英语,(2)当代汉语普通话,以及(3)英语的历史发展。我们的工作以概率框架语义为基础,在词汇创造力方面弥合了自然语言处理系统和人类之间的差距。
{"title":"Noun2Verb: Probabilistic Frame Semantics for Word Class Conversion","authors":"Lei Yu, Yang Xu","doi":"10.1162/coli_a_00447","DOIUrl":"https://doi.org/10.1162/coli_a_00447","url":null,"abstract":"Abstract Humans can flexibly extend word usages across different grammatical classes, a phenomenon known as word class conversion. Noun-to-verb conversion, or denominal verb (e.g., to Google a cheap flight), is one of the most prevalent forms of word class conversion. However, existing natural language processing systems are impoverished in interpreting and generating novel denominal verb usages. Previous work has suggested that novel denominal verb usages are comprehensible if the listener can compute the intended meaning based on shared knowledge with the speaker. Here we explore a computational formalism for this proposal couched in frame semantics. We present a formal framework, Noun2Verb, that simulates the production and comprehension of novel denominal verb usages by modeling shared knowledge of speaker and listener in semantic frames. We evaluate an incremental set of probabilistic models that learn to interpret and generate novel denominal verb usages via paraphrasing. We show that a model where the speaker and listener cooperatively learn the joint distribution over semantic frame elements better explains the empirical denominal verb usages than state-of-the-art language models, evaluated against data from (1) contemporary English in both adult and child speech, (2) contemporary Mandarin Chinese, and (3) the historical development of English. Our work grounds word class conversion in probabilistic frame semantics and bridges the gap between natural language processing systems and humans in lexical creativity.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47838054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
It Takes Two Flints to Make a Fire: Multitask Learning of Neural Relation and Explanation Classifiers 需要两个弗林特才能生火:神经关系和解释分类器的多任务学习
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-04-25 DOI: 10.1162/coli_a_00463
Zheng Tang, M. Surdeanu
We propose an explainable approach for relation extraction that mitigates the tension between generalization and explainability by jointly training for the two goals. Our approach uses a multi-task learning architecture, which jointly trains a classifier for relation extraction, and a sequence model that labels words in the context of the relations that explain the decisions of the relation classifier. We also convert the model outputs to rules to bring global explanations to this approach. This sequence model is trained using a hybrid strategy: supervised, when supervision from pre-existing patterns is available, and semi-supervised otherwise. In the latter situation, we treat the sequence model’s labels as latent variables, and learn the best assignment that maximizes the performance of the relation classifier. We evaluate the proposed approach on the two datasets and show that the sequence model provides labels that serve as accurate explanations for the relation classifier’s decisions, and, importantly, that the joint training generally improves the performance of the relation classifier. We also evaluate the performance of the generated rules and show that the new rules are a great add-on to the manual rules and bring the rule-based system much closer to the neural models.
我们提出了一种可解释的关系提取方法,通过为两个目标联合训练来缓解泛化和可解释性之间的紧张关系。我们的方法使用多任务学习架构,该架构联合训练用于关系提取的分类器,以及在解释关系分类器决策的关系上下文中标记单词的序列模型。我们还将模型输出转换为规则,以对这种方法进行全局解释。该序列模型使用混合策略进行训练:当可以从预先存在的模式进行监督时,进行监督,否则进行半监督。在后一种情况下,我们将序列模型的标签视为潜在变量,并学习最大化关系分类器性能的最佳分配。我们在两个数据集上评估了所提出的方法,并表明序列模型提供的标签可以作为关系分类器决策的准确解释,重要的是,联合训练通常可以提高关系分类器的性能。我们还评估了生成的规则的性能,并表明新规则是手动规则的一个很好的附加项,使基于规则的系统更接近神经模型。
{"title":"It Takes Two Flints to Make a Fire: Multitask Learning of Neural Relation and Explanation Classifiers","authors":"Zheng Tang, M. Surdeanu","doi":"10.1162/coli_a_00463","DOIUrl":"https://doi.org/10.1162/coli_a_00463","url":null,"abstract":"We propose an explainable approach for relation extraction that mitigates the tension between generalization and explainability by jointly training for the two goals. Our approach uses a multi-task learning architecture, which jointly trains a classifier for relation extraction, and a sequence model that labels words in the context of the relations that explain the decisions of the relation classifier. We also convert the model outputs to rules to bring global explanations to this approach. This sequence model is trained using a hybrid strategy: supervised, when supervision from pre-existing patterns is available, and semi-supervised otherwise. In the latter situation, we treat the sequence model’s labels as latent variables, and learn the best assignment that maximizes the performance of the relation classifier. We evaluate the proposed approach on the two datasets and show that the sequence model provides labels that serve as accurate explanations for the relation classifier’s decisions, and, importantly, that the joint training generally improves the performance of the relation classifier. We also evaluate the performance of the generated rules and show that the new rules are a great add-on to the manual rules and bring the rule-based system much closer to the neural models.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45265529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Revise and Resubmit: An Intertextual Model of Text-based Collaboration in Peer Review 修订和重新提交:同行评审中基于文本的协作的文本间模型
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-04-22 DOI: 10.1162/coli_a_00455
Ilia Kuznetsov, Jan Buchmann, Max Eichler, Iryna Gurevych
Abstract Peer review is a key component of the publishing process in most fields of science. Increasing submission rates put a strain on reviewing quality and efficiency, motivating the development of applications to support the reviewing and editorial work. While existing NLP studies focus on the analysis of individual texts, editorial assistance often requires modeling interactions between pairs of texts—yet general frameworks and datasets to support this scenario are missing. Relationships between texts are the core object of the intertextuality theory—a family of approaches in literary studies not yet operationalized in NLP. Inspired by prior theoretical work, we propose the first intertextual model of text-based collaboration, which encompasses three major phenomena that make up a full iteration of the review–revise–and–resubmit cycle: pragmatic tagging, linking, and long-document version alignment. While peer review is used across the fields of science and publication formats, existing datasets solely focus on conference-style review in computer science. Addressing this, we instantiate our proposed model in the first annotated multidomain corpus in journal-style post-publication open peer review, and provide detailed insights into the practical aspects of intertextual annotation. Our resource is a major step toward multidomain, fine-grained applications of NLP in editorial support for peer review, and our intertextual framework paves the path for general-purpose modeling of text-based collaboration. We make our corpus, detailed annotation guidelines, and accompanying code publicly available.1
摘要同行评审是大多数科学领域出版过程的关键组成部分。提交率的提高给审查质量和效率带来了压力,促使开发应用程序来支持审查和编辑工作。虽然现有的NLP研究侧重于对单个文本的分析,但编辑辅助通常需要对文本对之间的交互进行建模——然而,支持这种情况的通用框架和数据集却缺失了。文本之间的关系是互文性理论的核心对象,这是文学研究中尚未在NLP中实施的一系列方法。受先前理论工作的启发,我们提出了第一个基于文本的协作互文模型,该模型包括三个主要现象,构成了审查-修订-重新提交周期的完整迭代:语用标记、链接和长文档版本对齐。虽然同行评审在科学和出版格式领域都有使用,但现有的数据集只关注计算机科学中的会议式评审。针对这一点,我们在期刊风格的出版后开放同行评审中的第一个注释多领域语料库中实例化了我们提出的模型,并对互文注释的实际方面提供了详细的见解。我们的资源是NLP在编辑支持同行评审方面向多领域、细粒度应用迈出的重要一步,我们的互文框架为基于文本的协作的通用建模铺平了道路。我们公开我们的语料库、详细的注释指南和附带的代码。1
{"title":"Revise and Resubmit: An Intertextual Model of Text-based Collaboration in Peer Review","authors":"Ilia Kuznetsov, Jan Buchmann, Max Eichler, Iryna Gurevych","doi":"10.1162/coli_a_00455","DOIUrl":"https://doi.org/10.1162/coli_a_00455","url":null,"abstract":"Abstract Peer review is a key component of the publishing process in most fields of science. Increasing submission rates put a strain on reviewing quality and efficiency, motivating the development of applications to support the reviewing and editorial work. While existing NLP studies focus on the analysis of individual texts, editorial assistance often requires modeling interactions between pairs of texts—yet general frameworks and datasets to support this scenario are missing. Relationships between texts are the core object of the intertextuality theory—a family of approaches in literary studies not yet operationalized in NLP. Inspired by prior theoretical work, we propose the first intertextual model of text-based collaboration, which encompasses three major phenomena that make up a full iteration of the review–revise–and–resubmit cycle: pragmatic tagging, linking, and long-document version alignment. While peer review is used across the fields of science and publication formats, existing datasets solely focus on conference-style review in computer science. Addressing this, we instantiate our proposed model in the first annotated multidomain corpus in journal-style post-publication open peer review, and provide detailed insights into the practical aspects of intertextual annotation. Our resource is a major step toward multidomain, fine-grained applications of NLP in editorial support for peer review, and our intertextual framework paves the path for general-purpose modeling of text-based collaboration. We make our corpus, detailed annotation guidelines, and accompanying code publicly available.1","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49225007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Investigating Language Relationships in Multilingual Sentence Encoders Through the Lens of Linguistic Typology 语言类型学视角下的多语言句子编码器语言关系研究
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-04-13 DOI: 10.1162/coli_a_00444
Rochelle Choenni, Ekaterina Shutova
Abstract Multilingual sentence encoders have seen much success in cross-lingual model transfer for downstream NLP tasks. The success of this transfer is, however, dependent on the model’s ability to encode the patterns of cross-lingual similarity and variation. Yet, we know relatively little about the properties of individual languages or the general patterns of linguistic variation that the models encode. In this article, we investigate these questions by leveraging knowledge from the field of linguistic typology, which studies and documents structural and semantic variation across languages. We propose methods for separating language-specific subspaces within state-of-the-art multilingual sentence encoders (LASER, M-BERT, XLM, and XLM-R) with respect to a range of typological properties pertaining to lexical, morphological, and syntactic structure. Moreover, we investigate how typological information about languages is distributed across all layers of the models. Our results show interesting differences in encoding linguistic variation associated with different pretraining strategies. In addition, we propose a simple method to study how shared typological properties of languages are encoded in two state-of-the-art multilingual models—M-BERT and XLM-R. The results provide insight into their information-sharing mechanisms and suggest that these linguistic properties are encoded jointly across typologically similar languages in these models.
摘要多语言句子编码器在下游NLP任务的跨语言模型转移方面取得了很大成功。然而,这种转移的成功取决于模型对跨语言相似性和变异模式进行编码的能力。然而,我们对个别语言的特性或模型编码的语言变异的一般模式知之甚少。在这篇文章中,我们通过利用语言类型学领域的知识来研究这些问题,该领域研究并记录了不同语言之间的结构和语义变化。我们提出了在最先进的多语言句子编码器(LASER、M-BERT、XLM和XLM-R)中根据与词汇、形态和句法结构有关的一系列类型学特性来分离语言特定子空间的方法。此外,我们还研究了关于语言的类型信息是如何分布在模型的所有层中的。我们的研究结果显示,不同的预训练策略在编码语言变异方面存在着有趣的差异。此外,我们提出了一种简单的方法来研究语言的共享类型学属性是如何在两个最先进的多语言模型中编码的——M-BERT和XLM-R。研究结果深入了解了它们的信息共享机制,并表明这些语言属性是在这些模型中跨类型相似的语言共同编码的。
{"title":"Investigating Language Relationships in Multilingual Sentence Encoders Through the Lens of Linguistic Typology","authors":"Rochelle Choenni, Ekaterina Shutova","doi":"10.1162/coli_a_00444","DOIUrl":"https://doi.org/10.1162/coli_a_00444","url":null,"abstract":"Abstract Multilingual sentence encoders have seen much success in cross-lingual model transfer for downstream NLP tasks. The success of this transfer is, however, dependent on the model’s ability to encode the patterns of cross-lingual similarity and variation. Yet, we know relatively little about the properties of individual languages or the general patterns of linguistic variation that the models encode. In this article, we investigate these questions by leveraging knowledge from the field of linguistic typology, which studies and documents structural and semantic variation across languages. We propose methods for separating language-specific subspaces within state-of-the-art multilingual sentence encoders (LASER, M-BERT, XLM, and XLM-R) with respect to a range of typological properties pertaining to lexical, morphological, and syntactic structure. Moreover, we investigate how typological information about languages is distributed across all layers of the models. Our results show interesting differences in encoding linguistic variation associated with different pretraining strategies. In addition, we propose a simple method to study how shared typological properties of languages are encoded in two state-of-the-art multilingual models—M-BERT and XLM-R. The results provide insight into their information-sharing mechanisms and suggest that these linguistic properties are encoded jointly across typologically similar languages in these models.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-04-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44534157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
The Impact of Edge Displacement Vaserstein Distance on UD Parsing Performance 边缘位移、维瑟斯坦距离对UD解析性能的影响
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-04-07 DOI: 10.1162/coli_a_00440
Mark Anderson, Carlos Gómez-Rodríguez
Abstract We contribute to the discussion on parsing performance in NLP by introducing a measurement that evaluates the differences between the distributions of edge displacement (the directed distance of edges) seen in training and test data. We hypothesize that this measurement will be related to differences observed in parsing performance across treebanks. We motivate this by building upon previous work and then attempt to falsify this hypothesis by using a number of statistical methods. We establish that there is a statistical correlation between this measurement and parsing performance even when controlling for potential covariants. We then use this to establish a sampling technique that gives us an adversarial and complementary split. This gives an idea of the lower and upper bounds of parsing systems for a given treebank in lieu of freshly sampled data. In a broader sense, the methodology presented here can act as a reference for future correlation-based exploratory work in NLP.
摘要我们通过引入一种测量方法来评估训练数据和测试数据中的边缘位移(边缘的定向距离)分布之间的差异,为NLP中解析性能的讨论做出了贡献。我们假设这种测量将与在树库之间观察到的解析性能差异有关。我们通过在先前工作的基础上再接再厉,试图通过使用多种统计方法来证伪这一假设。我们确定,即使在控制潜在协变量的情况下,这种测量和解析性能之间也存在统计相关性。然后,我们利用这一点来建立一种采样技术,为我们提供对抗性和互补性的划分。这给出了一个给定树库的解析系统的下界和上界的概念,以代替新采样的数据。从更广泛的意义上讲,本文提出的方法可以作为未来NLP中基于相关性的探索工作的参考。
{"title":"The Impact of Edge Displacement Vaserstein Distance on UD Parsing Performance","authors":"Mark Anderson, Carlos Gómez-Rodríguez","doi":"10.1162/coli_a_00440","DOIUrl":"https://doi.org/10.1162/coli_a_00440","url":null,"abstract":"Abstract We contribute to the discussion on parsing performance in NLP by introducing a measurement that evaluates the differences between the distributions of edge displacement (the directed distance of edges) seen in training and test data. We hypothesize that this measurement will be related to differences observed in parsing performance across treebanks. We motivate this by building upon previous work and then attempt to falsify this hypothesis by using a number of statistical methods. We establish that there is a statistical correlation between this measurement and parsing performance even when controlling for potential covariants. We then use this to establish a sampling technique that gives us an adversarial and complementary split. This gives an idea of the lower and upper bounds of parsing systems for a given treebank in lieu of freshly sampled data. In a broader sense, the methodology presented here can act as a reference for future correlation-based exploratory work in NLP.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44597713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tractable Parsing for CCGs of Bounded Degree 有界度CCG的可牵引解析
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-04-07 DOI: 10.1162/coli_a_00441
Lena Katharina Schiffer, Marco Kuhlmann, G. Satta
Abstract Unlike other mildly context-sensitive formalisms, Combinatory Categorial Grammar (CCG) cannot be parsed in polynomial time when the size of the grammar is taken into account. Refining this result, we show that the parsing complexity of CCG is exponential only in the maximum degree of composition. When that degree is fixed, parsing can be carried out in polynomial time. Our finding is interesting from a linguistic perspective because a bounded degree of composition has been suggested as a universal constraint on natural language grammar. Moreover, ours is the first complexity result for a version of CCG that includes substitution rules, which are used in practical grammars but have been ignored in theoretical work.
与其他轻度上下文敏感的语法形式不同,当考虑到语法的大小时,组合范畴语法(CCG)不能在多项式时间内解析。对这一结果进行改进,我们发现CCG的解析复杂度仅在最大组合度下呈指数增长。当该程度固定时,解析可以在多项式时间内进行。从语言学的角度来看,我们的发现很有趣,因为有界的构成程度被认为是自然语言语法的普遍约束。此外,我们的研究是第一个包含替换规则的CCG版本的复杂性结果,替换规则在实际语法中使用,但在理论工作中被忽略。
{"title":"Tractable Parsing for CCGs of Bounded Degree","authors":"Lena Katharina Schiffer, Marco Kuhlmann, G. Satta","doi":"10.1162/coli_a_00441","DOIUrl":"https://doi.org/10.1162/coli_a_00441","url":null,"abstract":"Abstract Unlike other mildly context-sensitive formalisms, Combinatory Categorial Grammar (CCG) cannot be parsed in polynomial time when the size of the grammar is taken into account. Refining this result, we show that the parsing complexity of CCG is exponential only in the maximum degree of composition. When that degree is fixed, parsing can be carried out in polynomial time. Our finding is interesting from a linguistic perspective because a bounded degree of composition has been suggested as a universal constraint on natural language grammar. Moreover, ours is the first complexity result for a version of CCG that includes substitution rules, which are used in practical grammars but have been ignored in theoretical work.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44178341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
UDapter: Typology-based Language Adapters for Multilingual Dependency Parsing and Sequence Labeling 用于多语言依赖解析和序列标记的基于类型学的语言适配器
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-04-07 DOI: 10.1162/coli_a_00443
A. Üstün, Arianna Bisazza, G. Bouma, G. van Noord
Abstract Recent advances in multilingual language modeling have brought the idea of a truly universal parser closer to reality. However, such models are still not immune to the “curse of multilinguality”: Cross-language interference and restrained model capacity remain major obstacles. To address this, we propose a novel language adaptation approach by introducing contextual language adapters to a multilingual parser. Contextual language adapters make it possible to learn adapters via language embeddings while sharing model parameters across languages based on contextual parameter generation. Moreover, our method allows for an easy but effective integration of existing linguistic typology features into the parsing model. Because not all typological features are available for every language, we further combine typological feature prediction with parsing in a multi-task model that achieves very competitive parsing performance without the need for an external prediction system for missing features. The resulting parser, UDapter, can be used for dependency parsing as well as sequence labeling tasks such as POS tagging, morphological tagging, and NER. In dependency parsing, it outperforms strong monolingual and multilingual baselines on the majority of both high-resource and low-resource (zero-shot) languages, showing the success of the proposed adaptation approach. In sequence labeling tasks, our parser surpasses the baseline on high resource languages, and performs very competitively in a zero-shot setting. Our in-depth analyses show that adapter generation via typological features of languages is key to this success.1
摘要多语言语言建模的最新进展使真正通用的解析器的想法更接近现实。然而,这些模型仍然不能免受“多语诅咒”的影响:跨语言干扰和受限的模型能力仍然是主要障碍。为了解决这个问题,我们提出了一种新的语言自适应方法,将上下文语言适配器引入到多语言解析器中。上下文语言适配器使通过语言嵌入学习适配器成为可能,同时基于上下文参数生成跨语言共享模型参数。此外,我们的方法允许将现有的语言类型学特征简单而有效地集成到解析模型中。由于并非所有类型学特征都适用于每种语言,我们在多任务模型中进一步将类型学特征预测与解析相结合,实现了极具竞争力的解析性能,而不需要外部预测系统来预测缺失的特征。由此产生的解析器UDapter可用于相关性解析以及序列标记任务,如POS标记、形态标记和NER。在依赖解析中,它在大多数高资源和低资源(零样本)语言上都优于强单语和多语言基线,表明了所提出的自适应方法的成功。在序列标记任务中,我们的解析器超过了高资源语言的基线,并且在零样本设置中表现非常有竞争力。我们的深入分析表明,通过语言的类型特征生成适配器是这一成功的关键。1
{"title":"UDapter: Typology-based Language Adapters for Multilingual Dependency Parsing and Sequence Labeling","authors":"A. Üstün, Arianna Bisazza, G. Bouma, G. van Noord","doi":"10.1162/coli_a_00443","DOIUrl":"https://doi.org/10.1162/coli_a_00443","url":null,"abstract":"Abstract Recent advances in multilingual language modeling have brought the idea of a truly universal parser closer to reality. However, such models are still not immune to the “curse of multilinguality”: Cross-language interference and restrained model capacity remain major obstacles. To address this, we propose a novel language adaptation approach by introducing contextual language adapters to a multilingual parser. Contextual language adapters make it possible to learn adapters via language embeddings while sharing model parameters across languages based on contextual parameter generation. Moreover, our method allows for an easy but effective integration of existing linguistic typology features into the parsing model. Because not all typological features are available for every language, we further combine typological feature prediction with parsing in a multi-task model that achieves very competitive parsing performance without the need for an external prediction system for missing features. The resulting parser, UDapter, can be used for dependency parsing as well as sequence labeling tasks such as POS tagging, morphological tagging, and NER. In dependency parsing, it outperforms strong monolingual and multilingual baselines on the majority of both high-resource and low-resource (zero-shot) languages, showing the success of the proposed adaptation approach. In sequence labeling tasks, our parser surpasses the baseline on high resource languages, and performs very competitively in a zero-shot setting. Our in-depth analyses show that adapter generation via typological features of languages is key to this success.1","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46724233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Erratum for “Formal Basis of a Language Universal” “通用语言的形式基础”勘误
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-04-01 DOI: 10.1162/coli_x_00432
Miloš Stanojević, Mark Steedman
In the paper “Formal Basis of a Language Universal” by Miloš Stanojević and Mark Steedman in Computational Linguistics 47:1 (https://doi.org/10.1162/coli a 00394), there is an error in example (12) on page 17. The two occurrences of the notation W should appear as |W. The paper has been updated so that the paragraph reads: In the full theory, these rules are generalized to “second level” cases, in which the secondary function is of the form (Y|Z)|W such as the following “forward crossing” instance, in which — matches either / or in both input and output:
在MilošStanojević和Mark Steedman在计算语言学中的论文“语言普遍性的形式基础”47:1(https://doi.org/10.1162/colia 00394),在第17页的示例(12)中存在错误。符号W的两次出现应显示为|W。这篇论文已经更新,段落如下:在完整的理论中,这些规则被推广到“二级”情况,其中二阶函数的形式为(Y|Z)|W,例如以下“前向交叉”实例,其中--在输入和输出中都匹配/或:
{"title":"Erratum for “Formal Basis of a Language Universal”","authors":"Miloš Stanojević, Mark Steedman","doi":"10.1162/coli_x_00432","DOIUrl":"https://doi.org/10.1162/coli_x_00432","url":null,"abstract":"In the paper “Formal Basis of a Language Universal” by Miloš Stanojević and Mark Steedman in Computational Linguistics 47:1 (https://doi.org/10.1162/coli a 00394), there is an error in example (12) on page 17. The two occurrences of the notation W should appear as |W. The paper has been updated so that the paragraph reads: In the full theory, these rules are generalized to “second level” cases, in which the secondary function is of the form (Y|Z)|W such as the following “forward crossing” instance, in which — matches either / or in both input and output:","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43486180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Onception: Active Learning with Expert Advice for Real World Machine Translation Onception:现实世界机器翻译的专家建议下的主动学习
IF 9.3 2区 计算机科学 Q1 Arts and Humanities Pub Date : 2022-03-09 DOI: 10.1162/coli_a_00473
Vania Mendoncca, Ricardo Rei, Luísa Coheur, Alberto Sardinha INESC-ID Lisboa, Instituto Superior T'ecnico, AI Unbabel
Active learning can play an important role in low-resource settings (i.e., where annotated data is scarce), by selecting which instances may be more worthy to annotate. Most active learning approaches for Machine Translation assume the existence of a pool of sentences in a source language, and rely on human annotators to provide translations or post-edits, which can still be costly. In this article, we apply active learning to a real-world human-in-the-loop scenario in which we assume that: (1) the source sentences may not be readily available, but instead arrive in a stream; (2) the automatic translations receive feedback in the form of a rating, instead of a correct/edited translation, since the human-in-the-loop might be a user looking for a translation, but not be able to provide one. To tackle the challenge of deciding whether each incoming pair source–translations is worthy to query for human feedback, we resort to a number of stream-based active learning query strategies. Moreover, because we do not know in advance which query strategy will be the most adequate for a certain language pair and set of Machine Translation models, we propose to dynamically combine multiple strategies using prediction with expert advice. Our experiments on different language pairs and feedback settings show that using active learning allows us to converge on the best Machine Translation systems with fewer human interactions. Furthermore, combining multiple strategies using prediction with expert advice outperforms several individual active learning strategies with even fewer interactions, particularly in partial feedback settings.
主动学习可以通过选择哪些实例更值得注释,在低资源环境中(即注释数据稀缺的情况下)发挥重要作用。机器翻译的大多数主动学习方法都假设源语言中存在一组句子,并依靠人工注释器提供翻译或后期编辑,这仍然可能代价高昂。在本文中,我们将主动学习应用于现实世界中的人在循环场景,在该场景中,我们假设:(1)源句子可能不容易获得,而是以流的形式到达;(2) 自动翻译接收评级形式的反馈,而不是正确/编辑的翻译,因为循环中的人可能是寻找翻译但不能提供翻译的用户。为了解决决定每个输入对源翻译是否值得查询以获得人类反馈的挑战,我们采用了许多基于流的主动学习查询策略。此外,由于我们事先不知道哪种查询策略最适合某一语言对和机器翻译模型集,我们建议使用预测和专家建议动态组合多种策略。我们在不同语言对和反馈设置上的实验表明,使用主动学习可以让我们在更少的人机交互的情况下汇聚到最好的机器翻译系统上。此外,使用预测和专家建议的多种策略相结合,优于几种互动更少的个人主动学习策略,尤其是在部分反馈环境中。
{"title":"Onception: Active Learning with Expert Advice for Real World Machine Translation","authors":"Vania Mendoncca, Ricardo Rei, Luísa Coheur, Alberto Sardinha INESC-ID Lisboa, Instituto Superior T'ecnico, AI Unbabel","doi":"10.1162/coli_a_00473","DOIUrl":"https://doi.org/10.1162/coli_a_00473","url":null,"abstract":"Active learning can play an important role in low-resource settings (i.e., where annotated data is scarce), by selecting which instances may be more worthy to annotate. Most active learning approaches for Machine Translation assume the existence of a pool of sentences in a source language, and rely on human annotators to provide translations or post-edits, which can still be costly. In this article, we apply active learning to a real-world human-in-the-loop scenario in which we assume that: (1) the source sentences may not be readily available, but instead arrive in a stream; (2) the automatic translations receive feedback in the form of a rating, instead of a correct/edited translation, since the human-in-the-loop might be a user looking for a translation, but not be able to provide one. To tackle the challenge of deciding whether each incoming pair source–translations is worthy to query for human feedback, we resort to a number of stream-based active learning query strategies. Moreover, because we do not know in advance which query strategy will be the most adequate for a certain language pair and set of Machine Translation models, we propose to dynamically combine multiple strategies using prediction with expert advice. Our experiments on different language pairs and feedback settings show that using active learning allows us to converge on the best Machine Translation systems with fewer human interactions. Furthermore, combining multiple strategies using prediction with expert advice outperforms several individual active learning strategies with even fewer interactions, particularly in partial feedback settings.","PeriodicalId":55229,"journal":{"name":"Computational Linguistics","volume":null,"pages":null},"PeriodicalIF":9.3,"publicationDate":"2022-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49326100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Computational Linguistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1