首页 > 最新文献

Natural Language Engineering最新文献

英文 中文
Killing me softly: Creative and cognitive aspects of implicitness in abusive language online 温柔地杀死我:网络辱骂性语言中隐含的创造性和认知方面
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-08-03 DOI: 10.1017/s1351324922000316
Simona Frenda, V. Patti, Paolo Rosso
Abusive language is becoming a problematic issue for our society. The spread of messages that reinforce social and cultural intolerance could have dangerous effects in victims’ life. State-of-the-art technologies are often effective on detecting explicit forms of abuse, leaving unidentified the utterances with very weak offensive language but a strong hurtful effect. Scholars have advanced theoretical and qualitative observations on specific indirect forms of abusive language that make it hard to be recognized automatically. In this work, we propose a battery of statistical and computational analyses able to support these considerations, with a focus on creative and cognitive aspects of the implicitness, in texts coming from different sources such as social media and news. We experiment with transformers, multi-task learning technique, and a set of linguistic features to reveal the elements involved in the implicit and explicit manifestations of abuses, providing a solid basis for computational applications.
辱骂性语言正在成为我们社会的一个问题。传播强化社会和文化不容忍的信息可能对受害者的生活产生危险影响。最先进的技术往往能有效地检测出明显的虐待形式,使那些攻击性很弱但伤害性很强的话语无法识别。学者们对具体的间接形式的辱骂性语言进行了先进的理论和定性观察,使其难以被自动识别。在这项工作中,我们提出了一系列能够支持这些考虑的统计和计算分析,重点关注来自不同来源(如社交媒体和新闻)的文本中隐含性的创造性和认知方面。我们使用变形器、多任务学习技术和一组语言特征进行实验,以揭示滥用的隐性和显性表现所涉及的元素,为计算应用提供坚实的基础。
{"title":"Killing me softly: Creative and cognitive aspects of implicitness in abusive language online","authors":"Simona Frenda, V. Patti, Paolo Rosso","doi":"10.1017/s1351324922000316","DOIUrl":"https://doi.org/10.1017/s1351324922000316","url":null,"abstract":"\u0000 Abusive language is becoming a problematic issue for our society. The spread of messages that reinforce social and cultural intolerance could have dangerous effects in victims’ life. State-of-the-art technologies are often effective on detecting explicit forms of abuse, leaving unidentified the utterances with very weak offensive language but a strong hurtful effect. Scholars have advanced theoretical and qualitative observations on specific indirect forms of abusive language that make it hard to be recognized automatically. In this work, we propose a battery of statistical and computational analyses able to support these considerations, with a focus on creative and cognitive aspects of the implicitness, in texts coming from different sources such as social media and news. We experiment with transformers, multi-task learning technique, and a set of linguistic features to reveal the elements involved in the implicit and explicit manifestations of abuses, providing a solid basis for computational applications.","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":" ","pages":""},"PeriodicalIF":2.5,"publicationDate":"2022-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45738669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
An empirical study of incorporating syntactic constraints into BERT-based location metonymy resolution 将句法约束纳入基于BERT的位置转喻解析的实证研究
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-08-01 DOI: 10.1017/S135132492200033X
Hao Wang, Siyuan Du, X. Zheng, Li Meng
Abstract Metonymy resolution (MR) is a challenging task in the field of natural language processing. The task of MR aims to identify the metonymic usage of a word that employs an entity name to refer to another target entity. Recent BERT-based methods yield state-of-the-art performances. However, they neither make full use of the entity information nor explicitly consider syntactic structure. In contrast, in this paper, we argue that the metonymic process should be completed in a collaborative manner, relying on both lexical semantics and syntactic structure (syntax). This paper proposes a novel approach to enhancing BERT-based MR models with hard and soft syntactic constraints by using different types of convolutional neural networks to model dependency parse trees. Experimental results on benchmark datasets (e.g., ReLocaR, SemEval 2007 and WiMCor) confirm that leveraging syntactic information into fine pre-trained language models benefits MR tasks.
摘要转喻消解(MR)是自然语言处理领域中一项具有挑战性的任务。MR的任务旨在识别一个词的转喻用法,该词使用一个实体名称来指代另一个目标实体。最近基于BERT的方法产生了最先进的性能。然而,它们既没有充分利用实体信息,也没有明确考虑句法结构。相反,在本文中,我们认为转喻过程应该以合作的方式完成,同时依赖于词汇语义和句法结构(句法)。本文提出了一种新的方法,通过使用不同类型的卷积神经网络对依赖解析树进行建模,来增强具有硬和软句法约束的基于BERT的MR模型。在基准数据集(例如ReLocaR、SemEval 2007和WiMCor)上的实验结果证实,将句法信息用于精细的预训练语言模型有利于MR任务。
{"title":"An empirical study of incorporating syntactic constraints into BERT-based location metonymy resolution","authors":"Hao Wang, Siyuan Du, X. Zheng, Li Meng","doi":"10.1017/S135132492200033X","DOIUrl":"https://doi.org/10.1017/S135132492200033X","url":null,"abstract":"Abstract Metonymy resolution (MR) is a challenging task in the field of natural language processing. The task of MR aims to identify the metonymic usage of a word that employs an entity name to refer to another target entity. Recent BERT-based methods yield state-of-the-art performances. However, they neither make full use of the entity information nor explicitly consider syntactic structure. In contrast, in this paper, we argue that the metonymic process should be completed in a collaborative manner, relying on both lexical semantics and syntactic structure (syntax). This paper proposes a novel approach to enhancing BERT-based MR models with hard and soft syntactic constraints by using different types of convolutional neural networks to model dependency parse trees. Experimental results on benchmark datasets (e.g., ReLocaR, SemEval 2007 and WiMCor) confirm that leveraging syntactic information into fine pre-trained language models benefits MR tasks.","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":"29 1","pages":"669 - 692"},"PeriodicalIF":2.5,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47772219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From unified phrase representation to bilingual phrase alignment in an unsupervised manner 从统一短语表示到无监督的双语短语对齐
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-08-01 DOI: 10.1017/S1351324922000328
Jingshu Liu, E. Morin, Sebastian Peña Saldarriaga, Joseph Lark
Abstract Significant advances have been achieved in bilingual word-level alignment, yet the challenge remains for phrase-level alignment. Moreover, the need for parallel data is a critical drawback for the alignment task. This work proposes a system that alleviates these two problems: a unified phrase representation model using cross-lingual word embeddings as input and an unsupervised training algorithm inspired by recent works on neural machine translation. The system consists of a sequence-to-sequence architecture where a short sequence encoder constructs cross-lingual representations of phrases of any length, then an LSTM network decodes them w.r.t their contexts. After training with comparable corpora and existing key phrase extraction, our encoder provides cross-lingual phrase representations that can be compared without further transformation. Experiments on five data sets show that our method obtains state-of-the-art results on the bilingual phrase alignment task and improves the results of different length phrase alignment by a mean of 8.8 points in MAP.
摘要双语单词水平对齐取得了重大进展,但短语水平对齐仍然面临挑战。此外,对并行数据的需求是对齐任务的一个关键缺点。这项工作提出了一个缓解这两个问题的系统:一个使用跨语言单词嵌入作为输入的统一短语表示模型,以及一个受神经机器翻译最新工作启发的无监督训练算法。该系统由序列到序列结构组成,其中短序列编码器构建任何长度短语的跨语言表示,然后LSTM网络根据其上下文对其进行解码。在使用可比较的语料库和现有的关键短语提取进行训练后,我们的编码器提供了跨语言短语表示,无需进一步转换即可进行比较。在五个数据集上的实验表明,我们的方法在双语短语比对任务上获得了最先进的结果,并在MAP中将不同长度短语比对的结果平均提高了8.8分。
{"title":"From unified phrase representation to bilingual phrase alignment in an unsupervised manner","authors":"Jingshu Liu, E. Morin, Sebastian Peña Saldarriaga, Joseph Lark","doi":"10.1017/S1351324922000328","DOIUrl":"https://doi.org/10.1017/S1351324922000328","url":null,"abstract":"Abstract Significant advances have been achieved in bilingual word-level alignment, yet the challenge remains for phrase-level alignment. Moreover, the need for parallel data is a critical drawback for the alignment task. This work proposes a system that alleviates these two problems: a unified phrase representation model using cross-lingual word embeddings as input and an unsupervised training algorithm inspired by recent works on neural machine translation. The system consists of a sequence-to-sequence architecture where a short sequence encoder constructs cross-lingual representations of phrases of any length, then an LSTM network decodes them w.r.t their contexts. After training with comparable corpora and existing key phrase extraction, our encoder provides cross-lingual phrase representations that can be compared without further transformation. Experiments on five data sets show that our method obtains state-of-the-art results on the bilingual phrase alignment task and improves the results of different length phrase alignment by a mean of 8.8 points in MAP.","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":"29 1","pages":"643 - 668"},"PeriodicalIF":2.5,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48513181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Named-entity recognition in Turkish legal texts 土耳其法律文本中的命名实体识别
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-07-11 DOI: 10.1017/S1351324922000304
Can Çetindağ, Berkay Yazıcıoğlu, Aykut Koç
Abstract Natural language processing (NLP) technologies and applications in legal text processing are gaining momentum. Being one of the most prominent tasks in NLP, named-entity recognition (NER) can substantiate a great convenience for NLP in law due to the variety of named entities in the legal domain and their accentuated importance in legal documents. However, domain-specific NER models in the legal domain are not well studied. We present a NER model for Turkish legal texts with a custom-made corpus as well as several NER architectures based on conditional random fields and bidirectional long-short-term memories (BiLSTMs) to address the task. We also study several combinations of different word embeddings consisting of GloVe, Morph2Vec, and neural network-based character feature extraction techniques either with BiLSTM or convolutional neural networks. We report 92.27% F1 score with a hybrid word representation of GloVe and Morph2Vec with character-level features extracted with BiLSTM. Being an agglutinative language, the morphological structure of Turkish is also considered. To the best of our knowledge, our work is the first legal domain-specific NER study in Turkish and also the first study for an agglutinative language in the legal domain. Thus, our work can also have implications beyond the Turkish language.
摘要自然语言处理(NLP)技术及其在法律文本处理中的应用正在蓬勃发展。命名实体识别是NLP中最突出的任务之一,由于法律领域中命名实体的多样性及其在法律文件中的重要性,它可以为NLP在法律上提供极大的便利。然而,法律领域中特定领域的NER模型并没有得到很好的研究。我们提出了一个土耳其法律文本的NER模型,该模型具有定制的语料库,以及基于条件随机场和双向长短期记忆(BiLSTM)的几种NER架构,以解决该任务。我们还研究了不同单词嵌入的几种组合,包括GloVe、Morph2Vec和基于神经网络的字符特征提取技术,无论是使用BiLSTM还是卷积神经网络。我们报告了使用GloVe和Morph2Verc的混合词表示以及使用BiLSTM提取的字符级特征的92.27%的F1分数。土耳其语作为一种粘着语言,其形态结构也被认为是一种粘着性语言。据我们所知,我们的工作是第一次用土耳其语对特定法律领域的NER进行研究,也是第一次对法律领域中的粘性语言进行研究。因此,我们的工作也可能产生超出土耳其语的影响。
{"title":"Named-entity recognition in Turkish legal texts","authors":"Can Çetindağ, Berkay Yazıcıoğlu, Aykut Koç","doi":"10.1017/S1351324922000304","DOIUrl":"https://doi.org/10.1017/S1351324922000304","url":null,"abstract":"Abstract Natural language processing (NLP) technologies and applications in legal text processing are gaining momentum. Being one of the most prominent tasks in NLP, named-entity recognition (NER) can substantiate a great convenience for NLP in law due to the variety of named entities in the legal domain and their accentuated importance in legal documents. However, domain-specific NER models in the legal domain are not well studied. We present a NER model for Turkish legal texts with a custom-made corpus as well as several NER architectures based on conditional random fields and bidirectional long-short-term memories (BiLSTMs) to address the task. We also study several combinations of different word embeddings consisting of GloVe, Morph2Vec, and neural network-based character feature extraction techniques either with BiLSTM or convolutional neural networks. We report 92.27% F1 score with a hybrid word representation of GloVe and Morph2Vec with character-level features extracted with BiLSTM. Being an agglutinative language, the morphological structure of Turkish is also considered. To the best of our knowledge, our work is the first legal domain-specific NER study in Turkish and also the first study for an agglutinative language in the legal domain. Thus, our work can also have implications beyond the Turkish language.","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":"29 1","pages":"615 - 642"},"PeriodicalIF":2.5,"publicationDate":"2022-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46893283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Neural automated writing evaluation for Korean L2 writing 基于神经网络的韩语二语写作评价
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-07-07 DOI: 10.1017/S1351324922000298
Kyungtae Lim, Jayoung Song, Jungyeul Park
Abstract Although Korean language education is experiencing rapid growth in recent years and several studies have investigated automated writing evaluation (AWE) systems, AWE for Korean L2 writing still remains unexplored. Therefore, this study aims to develop and validate a state-of-the-art neural model AWE system which can be widely used for Korean language teaching and learning. Based on a Korean learner corpus, the proposed AWE is developed using natural language processing techniques such as part-of-speech tagging, syntactic parsing, and statistical language modeling to engineer linguistic features and a pre-trained neural language model. This study attempted to determine how neural network models use different linguistic features to improve AWE performance. Experimental results of the proposed AWE tool showed that the neural AWE system achieves high reliability for unseen test data from the corpus, which implies metrics used in the AWE system can help differentiate different proficiency levels and predict holistic scores. Furthermore, the results confirmed that the proposed linguistic features–syntactic complexity, quantitative complexity, and fluency–offer benefits that complement neural automated writing evaluation.
虽然近年来韩语教育发展迅速,一些研究也对自动写作评估(AWE)系统进行了研究,但对韩语第二语言写作的AWE系统仍未进行探索。因此,本研究旨在开发并验证一种可广泛应用于韩国语教与学的神经模型AWE系统。基于韩语学习者语料库,拟议的AWE使用自然语言处理技术(如词性标注、句法分析和统计语言建模)开发,以设计语言特征和预训练的神经语言模型。本研究试图确定神经网络模型如何使用不同的语言特征来提高AWE性能。实验结果表明,神经AWE系统对语料库中未见的测试数据具有较高的可靠性,这意味着AWE系统中使用的指标可以帮助区分不同的熟练程度并预测整体分数。此外,研究结果证实,提出的语言特征——句法复杂性、数量复杂性和流畅性——为神经自动写作评估提供了补充。
{"title":"Neural automated writing evaluation for Korean L2 writing","authors":"Kyungtae Lim, Jayoung Song, Jungyeul Park","doi":"10.1017/S1351324922000298","DOIUrl":"https://doi.org/10.1017/S1351324922000298","url":null,"abstract":"Abstract Although Korean language education is experiencing rapid growth in recent years and several studies have investigated automated writing evaluation (AWE) systems, AWE for Korean L2 writing still remains unexplored. Therefore, this study aims to develop and validate a state-of-the-art neural model AWE system which can be widely used for Korean language teaching and learning. Based on a Korean learner corpus, the proposed AWE is developed using natural language processing techniques such as part-of-speech tagging, syntactic parsing, and statistical language modeling to engineer linguistic features and a pre-trained neural language model. This study attempted to determine how neural network models use different linguistic features to improve AWE performance. Experimental results of the proposed AWE tool showed that the neural AWE system achieves high reliability for unseen test data from the corpus, which implies metrics used in the AWE system can help differentiate different proficiency levels and predict holistic scores. Furthermore, the results confirmed that the proposed linguistic features–syntactic complexity, quantitative complexity, and fluency–offer benefits that complement neural automated writing evaluation.","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":"29 1","pages":"1341 - 1363"},"PeriodicalIF":2.5,"publicationDate":"2022-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42622491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Construction Grammar Conceptual Network: Coordination-based graph method for semantic association analysis 构建语法概念网络:基于坐标的语义关联分析图方法
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-07-04 DOI: 10.1017/S1351324922000274
Benedikt Perak, Tajana Ban Kirigin
Abstract In this article, we present the Construction Grammar Conceptual Network method, developed for identifying lexical similarity and word sense discrimination in a syntactically tagged corpus, based on the cognitive linguistic assumption that coordination construction instantiates conceptual relatedness. This graph analysis method projects a semantic value onto a given coordinated syntactic dependency and constructs a second-order lexical network of lexical collocates with a high co-occurrence measure. The subsequent process of clustering and pruning the graph reveals lexical communities with high conceptual similarity, which are interpreted as associated senses of the source lexeme. We demonstrate the theory and its application to the task of identifying the conceptual structure and different meanings of nouns, adjectives and verbs using examples from different corpora, and explain the modulating effects of linguistic and graph parameters. This graph approach is based on syntactic dependency processing and can be used as a complementary method to other contemporary natural language processing resources to enrich semantic tasks such as word disambiguation, domain relatedness, sense structure, identification of synonymy, metonymy, and metaphoricity, as well as to automate comprehensive meta-reasoning about languages and identify cross/intra-cultural discourse variations of prototypical conceptualization patterns and knowledge representations. As a contribution, we provide a web-based app at http://emocnet.uniri.hr/.
摘要在这篇文章中,我们提出了构造语法概念网络方法,该方法是基于协调构造实例化概念关联的认知语言学假设,为识别句法标记语料库中的词汇相似性和词义辨别而开发的。该图分析方法将语义值投影到给定的协调句法依赖上,并构建具有高共现测度的词汇并置的二阶词汇网络。随后对图进行聚类和修剪的过程揭示了具有高度概念相似性的词汇群落,这些词汇群落被解释为源词位的关联意义。我们用不同语料库的例子展示了该理论及其在识别名词、形容词和动词的概念结构和不同含义方面的应用,并解释了语言和图形参数的调节作用。这种图方法基于句法依赖处理,可以作为当代其他自然语言处理资源的补充方法,丰富单词消歧、领域关联、意义结构、同义词识别、转喻和隐喻等语义任务,以及自动化关于语言的综合元推理,识别原型概念化模式和知识表征的跨文化/跨文化话语变体。作为贡献,我们在http://emocnet.uniri.hr/.
{"title":"Construction Grammar Conceptual Network: Coordination-based graph method for semantic association analysis","authors":"Benedikt Perak, Tajana Ban Kirigin","doi":"10.1017/S1351324922000274","DOIUrl":"https://doi.org/10.1017/S1351324922000274","url":null,"abstract":"Abstract In this article, we present the Construction Grammar Conceptual Network method, developed for identifying lexical similarity and word sense discrimination in a syntactically tagged corpus, based on the cognitive linguistic assumption that coordination construction instantiates conceptual relatedness. This graph analysis method projects a semantic value onto a given coordinated syntactic dependency and constructs a second-order lexical network of lexical collocates with a high co-occurrence measure. The subsequent process of clustering and pruning the graph reveals lexical communities with high conceptual similarity, which are interpreted as associated senses of the source lexeme. We demonstrate the theory and its application to the task of identifying the conceptual structure and different meanings of nouns, adjectives and verbs using examples from different corpora, and explain the modulating effects of linguistic and graph parameters. This graph approach is based on syntactic dependency processing and can be used as a complementary method to other contemporary natural language processing resources to enrich semantic tasks such as word disambiguation, domain relatedness, sense structure, identification of synonymy, metonymy, and metaphoricity, as well as to automate comprehensive meta-reasoning about languages and identify cross/intra-cultural discourse variations of prototypical conceptualization patterns and knowledge representations. As a contribution, we provide a web-based app at http://emocnet.uniri.hr/.","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":"29 1","pages":"584 - 614"},"PeriodicalIF":2.5,"publicationDate":"2022-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49089089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial fine-tuning tasks for yes/no question answering 人工微调是/否问题回答任务
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-06-30 DOI: 10.1017/s1351324922000286
Dimitris Dimitriadis, Grigorios Tsoumakas
Current research in yes/no question answering (QA) focuses on transfer learning techniques and transformer-based models. Models trained on large corpora are fine-tuned on tasks similar to yes/no QA, and then the captured knowledge is transferred for solving the yes/no QA task. Most previous studies use existing similar tasks, such as natural language inference or extractive QA, for the fine-tuning step. This paper follows a different perspective, hypothesizing that an artificial yes/no task can transfer useful knowledge for improving the performance of yes/no QA. We introduce three such tasks for this purpose, by adapting three corresponding existing tasks: candidate answer validation, sentiment classification, and lexical simplification. Furthermore, we experimented with three different variations of the BERT model (BERT base, RoBERTa, and ALBERT). The results show that our hypothesis holds true for all artificial tasks, despite the small size of the corresponding datasets that are used for the fine-tuning process, the differences between these tasks, the decisions that we made to adapt the original ones, and the tasks’ simplicity. This gives an alternative perspective on how to deal with the yes/no QA problem, that is more creative, and at the same time more flexible, as it can exploit multiple other existing tasks and corresponding datasets to improve yes/no QA models.
目前对是/否问答(QA)的研究主要集中在迁移学习技术和基于变换器的模型上。在大型语料库上训练的模型在类似于是/否QA的任务上进行微调,然后转移捕获的知识来解决是/否的QA任务。大多数先前的研究都使用现有的类似任务,如自然语言推理或提取QA来进行微调步骤。本文从另一个角度出发,假设一个人工的是/否任务可以传递有用的知识来提高是/否QA的性能。为此,我们引入了三个这样的任务,通过调整三个相应的现有任务:候选答案验证、情感分类和词汇简化。此外,我们对BERT模型的三种不同变体(BERT基础、RoBERTa和ALBERT)进行了实验。结果表明,我们的假设适用于所有人工任务,尽管用于微调过程的相应数据集规模较小,这些任务之间存在差异,我们为适应原始任务而做出的决定以及任务的简单性。这为如何处理是/否QA问题提供了另一种视角,这更具创造性,同时也更灵活,因为它可以利用多个其他现有任务和相应的数据集来改进是/否的QA模型。
{"title":"Artificial fine-tuning tasks for yes/no question answering","authors":"Dimitris Dimitriadis, Grigorios Tsoumakas","doi":"10.1017/s1351324922000286","DOIUrl":"https://doi.org/10.1017/s1351324922000286","url":null,"abstract":"\u0000 Current research in yes/no question answering (QA) focuses on transfer learning techniques and transformer-based models. Models trained on large corpora are fine-tuned on tasks similar to yes/no QA, and then the captured knowledge is transferred for solving the yes/no QA task. Most previous studies use existing similar tasks, such as natural language inference or extractive QA, for the fine-tuning step. This paper follows a different perspective, hypothesizing that an artificial yes/no task can transfer useful knowledge for improving the performance of yes/no QA. We introduce three such tasks for this purpose, by adapting three corresponding existing tasks: candidate answer validation, sentiment classification, and lexical simplification. Furthermore, we experimented with three different variations of the BERT model (BERT base, RoBERTa, and ALBERT). The results show that our hypothesis holds true for all artificial tasks, despite the small size of the corresponding datasets that are used for the fine-tuning process, the differences between these tasks, the decisions that we made to adapt the original ones, and the tasks’ simplicity. This gives an alternative perspective on how to deal with the yes/no QA problem, that is more creative, and at the same time more flexible, as it can exploit multiple other existing tasks and corresponding datasets to improve yes/no QA models.","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":" ","pages":""},"PeriodicalIF":2.5,"publicationDate":"2022-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42369049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated hate speech detection and span extraction in underground hacking and extremist forums 自动仇恨言论检测和跨度提取地下黑客和极端主义论坛
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-06-20 DOI: 10.1017/S1351324922000262
Linda Zhou, Andrew Caines, Ildiko Pete, Alice Hutchings
Abstract Hate speech is any kind of communication that attacks a person or a group based on their characteristics, such as gender, religion and race. Due to the availability of online platforms where people can express their (hateful) opinions, the amount of hate speech is steadily increasing that often leads to offline hate crimes. This paper focuses on understanding and detecting hate speech in underground hacking and extremist forums where cybercriminals and extremists, respectively, communicate with each other, and some of them are associated with criminal activity. Moreover, due to the lengthy posts, it would be beneficial to identify the specific span of text containing hateful content in order to assist site moderators with the removal of hate speech. This paper describes a hate speech dataset composed of posts extracted from HackForums, an online hacking forum, and Stormfront and Incels.co, two extremist forums. We combined our dataset with a Twitter hate speech dataset to train a multi-platform classifier. Our evaluation shows that a classifier trained on multiple sources of data does not always improve the performance compared to a mono-platform classifier. Finally, this is the first work on extracting hate speech spans from longer texts. The paper fine-tunes BERT (Bidirectional Encoder Representations from Transformers) and adopts two approaches – span prediction and sequence labelling. Both approaches successfully extract hateful spans and achieve an F1-score of at least 69%.
仇恨言论是基于性别、宗教和种族等特征攻击个人或群体的任何一种传播方式。由于人们可以在网络平台上表达自己的(仇恨)观点,仇恨言论的数量正在稳步增加,这往往导致线下仇恨犯罪。本文的重点是理解和检测地下黑客和极端主义论坛中的仇恨言论,这些论坛分别是网络犯罪分子和极端主义分子相互交流的地方,其中一些与犯罪活动有关。此外,由于帖子很长,确定包含仇恨内容的文本的具体跨度将是有益的,以协助网站版主删除仇恨言论。本文描述了一个仇恨言论数据集,该数据集由从HackForums(一个在线黑客论坛)以及Stormfront和Incels中提取的帖子组成。两个极端主义论坛。我们将我们的数据集与Twitter的仇恨言论数据集结合起来训练一个多平台分类器。我们的评估表明,与单平台分类器相比,在多个数据源上训练的分类器并不总能提高性能。最后,这是第一个从较长的文本中提取仇恨言论跨度的工作。本文对BERT (Bidirectional Encoder Representations from Transformers)进行了微调,采用了跨度预测和序列标记两种方法。两种方法都成功地提取了可恨跨度,并获得了至少69%的f1分数。
{"title":"Automated hate speech detection and span extraction in underground hacking and extremist forums","authors":"Linda Zhou, Andrew Caines, Ildiko Pete, Alice Hutchings","doi":"10.1017/S1351324922000262","DOIUrl":"https://doi.org/10.1017/S1351324922000262","url":null,"abstract":"Abstract Hate speech is any kind of communication that attacks a person or a group based on their characteristics, such as gender, religion and race. Due to the availability of online platforms where people can express their (hateful) opinions, the amount of hate speech is steadily increasing that often leads to offline hate crimes. This paper focuses on understanding and detecting hate speech in underground hacking and extremist forums where cybercriminals and extremists, respectively, communicate with each other, and some of them are associated with criminal activity. Moreover, due to the lengthy posts, it would be beneficial to identify the specific span of text containing hateful content in order to assist site moderators with the removal of hate speech. This paper describes a hate speech dataset composed of posts extracted from HackForums, an online hacking forum, and Stormfront and Incels.co, two extremist forums. We combined our dataset with a Twitter hate speech dataset to train a multi-platform classifier. Our evaluation shows that a classifier trained on multiple sources of data does not always improve the performance compared to a mono-platform classifier. Finally, this is the first work on extracting hate speech spans from longer texts. The paper fine-tunes BERT (Bidirectional Encoder Representations from Transformers) and adopts two approaches – span prediction and sequence labelling. Both approaches successfully extract hateful spans and achieve an F1-score of at least 69%.","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":"29 1","pages":"1247 - 1274"},"PeriodicalIF":2.5,"publicationDate":"2022-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44662755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
NLE volume 28 issue 4 Cover and Back matter NLE第28卷第4期封面和封底
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-06-16 DOI: 10.1017/s1351324922000250
{"title":"NLE volume 28 issue 4 Cover and Back matter","authors":"","doi":"10.1017/s1351324922000250","DOIUrl":"https://doi.org/10.1017/s1351324922000250","url":null,"abstract":"","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":" ","pages":"b1 - b2"},"PeriodicalIF":2.5,"publicationDate":"2022-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43592862","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NLE volume 28 issue 4 Cover and Front matter NLE第28卷第4期封面和封面问题
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-06-16 DOI: 10.1017/s1351324922000249
R. Mitkov, B. Boguraev
{"title":"NLE volume 28 issue 4 Cover and Front matter","authors":"R. Mitkov, B. Boguraev","doi":"10.1017/s1351324922000249","DOIUrl":"https://doi.org/10.1017/s1351324922000249","url":null,"abstract":"","PeriodicalId":49143,"journal":{"name":"Natural Language Engineering","volume":" ","pages":"f1 - f2"},"PeriodicalIF":2.5,"publicationDate":"2022-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43482558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Natural Language Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1