首页 > 最新文献

Recent Advances in Natural Language Processing最新文献

英文 中文
Exploring the Landscape of Natural Language Processing Research 探索自然语言处理研究的前景
Pub Date : 2023-07-20 DOI: 10.26615/978-954-452-092-2_111
Tim Schopf, Karim Arabi, F. Matthes
As an efficient approach to understand, generate, and process natural language texts, research in natural language processing (NLP) has exhibited a rapid spread and wide adoption in recent years. Given the increasing research work in this area, several NLP-related approaches have been surveyed in the research community. However, a comprehensive study that categorizes established topics, identifies trends, and outlines areas for future research remains absent. Contributing to closing this gap, we have systematically classified and analyzed research papers in the ACL Anthology. As a result, we present a structured overview of the research landscape, provide a taxonomy of fields of study in NLP, analyze recent developments in NLP, summarize our findings, and highlight directions for future work.
作为一种理解、生成和处理自然语言文本的有效方法,自然语言处理(NLP)研究近年来得到了迅速推广和广泛应用。鉴于该领域的研究工作日益增多,研究界已对几种与 NLP 相关的方法进行了调查。然而,目前仍缺乏一项全面的研究,对已确立的主题进行分类、确定趋势并勾勒出未来研究的领域。为了填补这一空白,我们对《ACL 文选》中的研究论文进行了系统的分类和分析。因此,我们对研究领域进行了结构化概述,对 NLP 的研究领域进行了分类,分析了 NLP 的最新发展,总结了我们的发现,并强调了未来的工作方向。
{"title":"Exploring the Landscape of Natural Language Processing Research","authors":"Tim Schopf, Karim Arabi, F. Matthes","doi":"10.26615/978-954-452-092-2_111","DOIUrl":"https://doi.org/10.26615/978-954-452-092-2_111","url":null,"abstract":"As an efficient approach to understand, generate, and process natural language texts, research in natural language processing (NLP) has exhibited a rapid spread and wide adoption in recent years. Given the increasing research work in this area, several NLP-related approaches have been surveyed in the research community. However, a comprehensive study that categorizes established topics, identifies trends, and outlines areas for future research remains absent. Contributing to closing this gap, we have systematically classified and analyzed research papers in the ACL Anthology. As a result, we present a structured overview of the research landscape, provide a taxonomy of fields of study in NLP, analyze recent developments in NLP, summarize our findings, and highlight directions for future work.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"80 1","pages":"1034-1045"},"PeriodicalIF":0.0,"publicationDate":"2023-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139356984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AspectCSE: Sentence Embeddings for Aspect-Based Semantic Textual Similarity Using Contrastive Learning and Structured Knowledge AspectCSE:利用对比学习和结构化知识实现基于方面的语义文本相似性的句子嵌入
Pub Date : 2023-07-15 DOI: 10.26615/978-954-452-092-2_113
Tim Schopf, Emanuel Gerber, Malte Ostendorff, F. Matthes
Generic sentence embeddings provide coarse-grained approximation of semantic textual similarity, but ignore specific aspects that make texts similar. Conversely, aspect-based sentence embeddings provide similarities between texts based on certain predefined aspects. Thus, similarity predictions of texts are more targeted to specific requirements and more easily explainable. In this paper, we present AspectCSE, an approach for aspect-based contrastive learning of sentence embeddings. Results indicate that AspectCSE achieves an average improvement of 3.97% on information retrieval tasks across multiple aspects compared to the previous best results. We also propose the use of Wikidata knowledge graph properties to train models of multi-aspect sentence embeddings in which multiple specific aspects are simultaneously considered during similarity predictions. We demonstrate that multi-aspect embeddings outperform even single-aspect embeddings on aspect-specific information retrieval tasks. Finally, we examine the aspect-based sentence embedding space and demonstrate that embeddings of semantically similar aspect labels are often close, even without explicit similarity training between different aspect labels.
通用句子嵌入提供了语义文本相似性的粗粒度近似值,但忽略了使文本相似的特定方面。相反,基于方面的句子嵌入则根据某些预定义的方面提供文本之间的相似性。因此,文本的相似性预测更能满足特定要求,也更容易解释。本文介绍了基于方面的句子嵌入对比学习方法 AspectCSE。结果表明,与之前的最佳结果相比,AspectCSE 在多个方面的信息检索任务中平均提高了 3.97%。我们还建议使用维基数据知识图谱属性来训练多方面句子嵌入模型,其中在进行相似性预测时会同时考虑多个特定方面。我们证明,在特定方面的信息检索任务中,多方面嵌入甚至优于单方面嵌入。最后,我们研究了基于方面的句子嵌入空间,并证明即使不同方面标签之间没有明确的相似性训练,语义上相似的方面标签的嵌入通常也很接近。
{"title":"AspectCSE: Sentence Embeddings for Aspect-Based Semantic Textual Similarity Using Contrastive Learning and Structured Knowledge","authors":"Tim Schopf, Emanuel Gerber, Malte Ostendorff, F. Matthes","doi":"10.26615/978-954-452-092-2_113","DOIUrl":"https://doi.org/10.26615/978-954-452-092-2_113","url":null,"abstract":"Generic sentence embeddings provide coarse-grained approximation of semantic textual similarity, but ignore specific aspects that make texts similar. Conversely, aspect-based sentence embeddings provide similarities between texts based on certain predefined aspects. Thus, similarity predictions of texts are more targeted to specific requirements and more easily explainable. In this paper, we present AspectCSE, an approach for aspect-based contrastive learning of sentence embeddings. Results indicate that AspectCSE achieves an average improvement of 3.97% on information retrieval tasks across multiple aspects compared to the previous best results. We also propose the use of Wikidata knowledge graph properties to train models of multi-aspect sentence embeddings in which multiple specific aspects are simultaneously considered during similarity predictions. We demonstrate that multi-aspect embeddings outperform even single-aspect embeddings on aspect-specific information retrieval tasks. Finally, we examine the aspect-based sentence embedding space and demonstrate that embeddings of semantically similar aspect labels are often close, even without explicit similarity training between different aspect labels.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"75 1","pages":"1054-1065"},"PeriodicalIF":0.0,"publicationDate":"2023-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139359074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Domain Adaptation of Sentence Embeddings Using Adapters 使用适配器对句子嵌入进行高效的领域适应性调整
Pub Date : 2023-07-06 DOI: 10.26615/978-954-452-092-2_112
Tim Schopf, Dennis Schneider, F. Matthes
Sentence embeddings enable us to capture the semantic similarity of short texts. Most sentence embedding models are trained for general semantic textual similarity tasks. Therefore, to use sentence embeddings in a particular domain, the model must be adapted to it in order to achieve good results. Usually, this is done by fine-tuning the entire sentence embedding model for the domain of interest. While this approach yields state-of-the-art results, all of the model’s weights are updated during fine-tuning, making this method resource-intensive. Therefore, instead of fine-tuning entire sentence embedding models for each target domain individually, we propose to train lightweight adapters. These domain-specific adapters do not require fine-tuning all underlying sentence embedding model parameters. Instead, we only train a small number of additional parameters while keeping the weights of the underlying sentence embedding model fixed. Training domain-specific adapters allows always using the same base model and only exchanging the domain-specific adapters to adapt sentence embeddings to a specific domain. We show that using adapters for parameter-efficient domain adaptation of sentence embeddings yields competitive performance within 1% of a domain-adapted, entirely fine-tuned sentence embedding model while only training approximately 3.6% of the parameters.
句子嵌入使我们能够捕捉短文的语义相似性。大多数句子嵌入模型都是针对一般语义文本相似性任务训练的。因此,要在特定领域使用句子嵌入,必须对模型进行调整,以取得良好的效果。通常,要做到这一点,需要针对感兴趣的领域对整个句子嵌入模型进行微调。虽然这种方法能获得最先进的结果,但在微调过程中,模型的所有权重都要更新,因此这种方法需要大量资源。因此,我们建议训练轻量级适配器,而不是为每个目标域单独微调整个句子嵌入模型。这些针对特定领域的适配器不需要微调所有底层句子嵌入模型参数。相反,我们只需训练少量附加参数,同时保持底层句子嵌入模型的权重固定不变。通过训练特定领域适配器,可以始终使用相同的基础模型,只需交换特定领域适配器即可将句子嵌入调整到特定领域。我们的研究表明,使用适配器对句子嵌入进行参数高效的领域适配,可以获得与完全微调的领域适配句子嵌入模型 1%以内的性能,而只需训练约 3.6% 的参数。
{"title":"Efficient Domain Adaptation of Sentence Embeddings Using Adapters","authors":"Tim Schopf, Dennis Schneider, F. Matthes","doi":"10.26615/978-954-452-092-2_112","DOIUrl":"https://doi.org/10.26615/978-954-452-092-2_112","url":null,"abstract":"Sentence embeddings enable us to capture the semantic similarity of short texts. Most sentence embedding models are trained for general semantic textual similarity tasks. Therefore, to use sentence embeddings in a particular domain, the model must be adapted to it in order to achieve good results. Usually, this is done by fine-tuning the entire sentence embedding model for the domain of interest. While this approach yields state-of-the-art results, all of the model’s weights are updated during fine-tuning, making this method resource-intensive. Therefore, instead of fine-tuning entire sentence embedding models for each target domain individually, we propose to train lightweight adapters. These domain-specific adapters do not require fine-tuning all underlying sentence embedding model parameters. Instead, we only train a small number of additional parameters while keeping the weights of the underlying sentence embedding model fixed. Training domain-specific adapters allows always using the same base model and only exchanging the domain-specific adapters to adapt sentence embeddings to a specific domain. We show that using adapters for parameter-efficient domain adaptation of sentence embeddings yields competitive performance within 1% of a domain-adapted, entirely fine-tuned sentence embedding model while only training approximately 3.6% of the parameters.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"1 1","pages":"1046-1053"},"PeriodicalIF":0.0,"publicationDate":"2023-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139362226","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural Machine Translation for Sinhala-English Code-Mixed Text 僧伽罗语-英语语码混合文本的神经机器翻译
Pub Date : 2022-12-30 DOI: 10.26615/978-954-452-072-4_082
Archchana Kugathasan, S. Sumathipala
Code-mixing has become a moving method of communication among multilingual speakers. Most of the social media content of the multilingual societies are written in code-mixed text. However, most of the current translation systems neglect to convert code-mixed texts to a standard language. Most of the user written code-mixed content in social media remains unprocessed due to the unavailability of linguistic resource such as parallel corpus. This paper proposes a Neural Machine Translation(NMT) model to translate the Sinhala-English code-mixed text to the Sinhala language. Due to the limited resources available for Sinhala-English code-mixed(SECM) text, a parallel corpus is created with SECM sentences and Sinhala sentences. Srilankan social media sites contain SECM texts more frequently than the standard languages. The model proposed for code-mixed text translation in this study is a combination of Encoder-Decoder framework with LSTM units and Teachers Forcing Algorithm. The translated sentences from the model are evaluated using BLEU(Bilingual Evaluation Understudy) metric. Our model achieved a remarkable BLEU score for the translation.
语码混合已经成为多语言使用者之间的一种交流方式。多语言社会的大多数社交媒体内容都是用代码混合文本编写的。然而,目前大多数翻译系统忽略了将代码混合文本转换为标准语言。由于平行语料库等语言资源的缺乏,社交媒体中大部分用户编写的代码混合内容都没有得到处理。本文提出了一种神经机器翻译(NMT)模型,用于将僧伽罗语-英语代码混合文本翻译成僧伽罗语。由于僧伽罗语-英语代码混合(SECM)文本的资源有限,我们创建了一个由SECM句子和僧伽罗语句子组成的平行语料库。斯里兰卡的社交媒体网站包含SECM文本的频率高于标准语言。本文提出的代码混合文本翻译模型是将编码器-解码器框架与LSTM单元和教师强制算法相结合。使用BLEU(双语评价替补)度量对模型中的翻译句子进行评估。我们的模型在翻译中取得了显著的BLEU分数。
{"title":"Neural Machine Translation for Sinhala-English Code-Mixed Text","authors":"Archchana Kugathasan, S. Sumathipala","doi":"10.26615/978-954-452-072-4_082","DOIUrl":"https://doi.org/10.26615/978-954-452-072-4_082","url":null,"abstract":"Code-mixing has become a moving method of communication among multilingual speakers. Most of the social media content of the multilingual societies are written in code-mixed text. However, most of the current translation systems neglect to convert code-mixed texts to a standard language. Most of the user written code-mixed content in social media remains unprocessed due to the unavailability of linguistic resource such as parallel corpus. This paper proposes a Neural Machine Translation(NMT) model to translate the Sinhala-English code-mixed text to the Sinhala language. Due to the limited resources available for Sinhala-English code-mixed(SECM) text, a parallel corpus is created with SECM sentences and Sinhala sentences. Srilankan social media sites contain SECM texts more frequently than the standard languages. The model proposed for code-mixed text translation in this study is a combination of Encoder-Decoder framework with LSTM units and Teachers Forcing Algorithm. The translated sentences from the model are evaluated using BLEU(Bilingual Evaluation Understudy) metric. Our model achieved a remarkable BLEU score for the translation.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123185786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Domain-Independent Holistic Approach to Deception Detection 欺骗检测的领域独立整体方法
Pub Date : 2021-11-21 DOI: 10.26615/978-954-452-072-4_147
Sadat Shahriar, Arjun Mukherjee, O. Gnawali
The deception in the text can be of different forms in different domains, including fake news, rumor tweets, and spam emails. Irrespective of the domain, the main intent of the deceptive text is to deceit the reader. Although domain-specific deception detection exists, domain-independent deception detection can provide a holistic picture, which can be crucial to understand how deception occurs in the text. In this paper, we detect deception in a domain-independent setting using deep learning architectures. Our method outperforms the State-of-the-Art performance of most benchmark datasets with an overall accuracy of 93.42% and F1-Score of 93.22%. The domain-independent training allows us to capture subtler nuances of deceptive writing style. Furthermore, we analyze how much in-domain data may be helpful to accurately detect deception, especially for the cases where data may not be readily available to train. Our results and analysis indicate that there may be a universal pattern of deception lying in-between the text independent of the domain, which can create a novel area of research and open up new avenues in the field of deception detection.
文本中的欺骗可以在不同的领域以不同的形式出现,包括假新闻、谣言推文和垃圾邮件。无论在哪个领域,欺骗性文本的主要目的都是欺骗读者。虽然存在特定领域的欺骗检测,但领域独立的欺骗检测可以提供一个整体的画面,这对于理解欺骗在文本中是如何发生的至关重要。在本文中,我们使用深度学习架构在领域独立的设置中检测欺骗。我们的方法优于大多数基准数据集的最先进性能,总体精度为93.42%,F1-Score为93.22%。独立于领域的训练使我们能够捕捉到欺骗性写作风格的细微差别。此外,我们分析了多少域内数据可能有助于准确检测欺骗,特别是在数据可能不容易获得训练的情况下。我们的结果和分析表明,在文本之间可能存在一种独立于域的普遍欺骗模式,这可以创造一个新的研究领域,并在欺骗检测领域开辟新的途径。
{"title":"A Domain-Independent Holistic Approach to Deception Detection","authors":"Sadat Shahriar, Arjun Mukherjee, O. Gnawali","doi":"10.26615/978-954-452-072-4_147","DOIUrl":"https://doi.org/10.26615/978-954-452-072-4_147","url":null,"abstract":"The deception in the text can be of different forms in different domains, including fake news, rumor tweets, and spam emails. Irrespective of the domain, the main intent of the deceptive text is to deceit the reader. Although domain-specific deception detection exists, domain-independent deception detection can provide a holistic picture, which can be crucial to understand how deception occurs in the text. In this paper, we detect deception in a domain-independent setting using deep learning architectures. Our method outperforms the State-of-the-Art performance of most benchmark datasets with an overall accuracy of 93.42% and F1-Score of 93.22%. The domain-independent training allows us to capture subtler nuances of deceptive writing style. Furthermore, we analyze how much in-domain data may be helpful to accurately detect deception, especially for the cases where data may not be readily available to train. Our results and analysis indicate that there may be a universal pattern of deception lying in-between the text independent of the domain, which can create a novel area of research and open up new avenues in the field of deception detection.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127731564","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Not All Comments Are Equal: Insights into Comment Moderation from a Topic-Aware Model 并非所有评论都是平等的:基于主题感知模型的评论审核洞察
Pub Date : 2021-09-21 DOI: 10.26615/978-954-452-072-4_185
Elaine Zosa, Ravi Shekhar, Mladen Karan, Matthew Purver
Moderation of reader comments is a significant problem for online news platforms. Here, we experiment with models for automatic moderation, using a dataset of comments from a popular Croatian newspaper. Our analysis shows that while comments that violate the moderation rules mostly share common linguistic and thematic features, their content varies across the different sections of the newspaper. We therefore make our models topic-aware, incorporating semantic features from a topic model into the classification decision. Our results show that topic information improves the performance of the model, increases its confidence in correct outputs, and helps us understand the model’s outputs.
读者评论的适度管理是网络新闻平台面临的一个重要问题。在这里,我们使用来自克罗地亚一家流行报纸的评论数据集,对自动审核模型进行了实验。我们的分析表明,虽然违反审查规则的评论大多具有共同的语言和主题特征,但它们的内容在报纸的不同部分有所不同。因此,我们使模型具有主题意识,将主题模型的语义特征合并到分类决策中。我们的结果表明,主题信息提高了模型的性能,增加了模型对正确输出的信心,并帮助我们理解模型的输出。
{"title":"Not All Comments Are Equal: Insights into Comment Moderation from a Topic-Aware Model","authors":"Elaine Zosa, Ravi Shekhar, Mladen Karan, Matthew Purver","doi":"10.26615/978-954-452-072-4_185","DOIUrl":"https://doi.org/10.26615/978-954-452-072-4_185","url":null,"abstract":"Moderation of reader comments is a significant problem for online news platforms. Here, we experiment with models for automatic moderation, using a dataset of comments from a popular Croatian newspaper. Our analysis shows that while comments that violate the moderation rules mostly share common linguistic and thematic features, their content varies across the different sections of the newspaper. We therefore make our models topic-aware, incorporating semantic features from a topic model into the classification decision. Our results show that topic information improves the performance of the model, increases its confidence in correct outputs, and helps us understand the model’s outputs.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122501742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Cross-lingual Offensive Language Identification for Low Resource Languages: The Case of Marathi 低资源语言的跨语攻击性语言识别:以马拉地语为例
Pub Date : 2021-09-08 DOI: 10.26615/978-954-452-072-4_050
Saurabh Gaikwad, Tharindu Ranasinghe, Marcos Zampieri, C. Homan
The widespread presence of offensive language on social media motivated the development of systems capable of recognizing such content automatically. Apart from a few notable exceptions, most research on automatic offensive language identification has dealt with English. To address this shortcoming, we introduce MOLD, the Marathi Offensive Language Dataset. MOLD is the first dataset of its kind compiled for Marathi, thus opening a new domain for research in low-resource Indo-Aryan languages. We present results from several machine learning experiments on this dataset, including zero-short and other transfer learning experiments on state-of-the-art cross-lingual transformers from existing data in Bengali, English, and Hindi.
社交媒体上广泛存在的攻击性语言促使开发能够自动识别此类内容的系统。除了少数值得注意的例外,大多数关于攻击性语言自动识别的研究都是针对英语的。为了解决这个缺点,我们引入了MOLD,马拉地攻击性语言数据集。MOLD是第一个为马拉地语编写的数据集,从而为低资源的印度雅利安语言的研究开辟了一个新的领域。我们展示了在该数据集上进行的几个机器学习实验的结果,包括对孟加拉语、英语和印地语现有数据进行的最先进的跨语言转换器的零短和其他迁移学习实验。
{"title":"Cross-lingual Offensive Language Identification for Low Resource Languages: The Case of Marathi","authors":"Saurabh Gaikwad, Tharindu Ranasinghe, Marcos Zampieri, C. Homan","doi":"10.26615/978-954-452-072-4_050","DOIUrl":"https://doi.org/10.26615/978-954-452-072-4_050","url":null,"abstract":"The widespread presence of offensive language on social media motivated the development of systems capable of recognizing such content automatically. Apart from a few notable exceptions, most research on automatic offensive language identification has dealt with English. To address this shortcoming, we introduce MOLD, the Marathi Offensive Language Dataset. MOLD is the first dataset of its kind compiled for Marathi, thus opening a new domain for research in low-resource Indo-Aryan languages. We present results from several machine learning experiments on this dataset, including zero-short and other transfer learning experiments on state-of-the-art cross-lingual transformers from existing data in Bengali, English, and Hindi.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115765937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
NEREL: A Russian Dataset with Nested Named Entities, Relations and Events NEREL:一个带有嵌套命名实体、关系和事件的俄语数据集
Pub Date : 2021-08-30 DOI: 10.26615/978-954-452-072-4_100
Natalia V. Loukachevitch, E. Artemova, Tatiana Batura, Pavel Braslavski, Ilia Denisov, V. Ivanov, S. Manandhar, Alexander Pugachev, E. Tutubalina
In this paper, we present NEREL, a Russian dataset for named entity recognition and relation extraction. NEREL is significantly larger than existing Russian datasets: to date it contains 56K annotated named entities and 39K annotated relations. Its important difference from previous datasets is annotation of nested named entities, as well as relations within nested entities and at the discourse level. NEREL can facilitate development of novel models that can extract relations between nested named entities, as well as relations on both sentence and document levels. NEREL also contains the annotation of events involving named entities and their roles in the events. The NEREL collection is available via https://github.com/nerel-ds/NEREL.
在本文中,我们提出了NEREL,一个用于命名实体识别和关系提取的俄语数据集。NEREL比现有的俄语数据集大得多:到目前为止,它包含56K个带注释的命名实体和39K个带注释的关系。它与以前的数据集的重要区别是对嵌套命名实体的注释,以及嵌套实体内部和话语级别的关系。NEREL可以促进新模型的开发,这些模型可以提取嵌套命名实体之间的关系,以及句子和文档级别上的关系。NEREL还包含涉及命名实体及其在事件中的角色的事件注释。NEREL系列可通过https://github.com/nerel-ds/NEREL获得。
{"title":"NEREL: A Russian Dataset with Nested Named Entities, Relations and Events","authors":"Natalia V. Loukachevitch, E. Artemova, Tatiana Batura, Pavel Braslavski, Ilia Denisov, V. Ivanov, S. Manandhar, Alexander Pugachev, E. Tutubalina","doi":"10.26615/978-954-452-072-4_100","DOIUrl":"https://doi.org/10.26615/978-954-452-072-4_100","url":null,"abstract":"In this paper, we present NEREL, a Russian dataset for named entity recognition and relation extraction. NEREL is significantly larger than existing Russian datasets: to date it contains 56K annotated named entities and 39K annotated relations. Its important difference from previous datasets is annotation of nested named entities, as well as relations within nested entities and at the discourse level. NEREL can facilitate development of novel models that can extract relations between nested named entities, as well as relations on both sentence and document levels. NEREL also contains the annotation of events involving named entities and their roles in the events. The NEREL collection is available via https://github.com/nerel-ds/NEREL.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"116 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134463849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Interpretable Propaganda Detection in News Articles 新闻文章中的可解释性宣传检测
Pub Date : 2021-08-29 DOI: 10.26615/978-954-452-072-4_179
Seunghak Yu, Giovanni Da San Martino, Mitra Mohtarami, James R. Glass, Preslav Nakov
Online users today are exposed to misleading and propagandistic news articles and media posts on a daily basis. To counter thus, a number of approaches have been designed aiming to achieve a healthier and safer online news and media consumption. Automatic systems are able to support humans in detecting such content; yet, a major impediment to their broad adoption is that besides being accurate, the decisions of such systems need also to be interpretable in order to be trusted and widely adopted by users. Since misleading and propagandistic content influences readers through the use of a number of deception techniques, we propose to detect and to show the use of such techniques as a way to offer interpretability. In particular, we define qualitatively descriptive features and we analyze their suitability for detecting deception techniques. We further show that our interpretable features can be easily combined with pre-trained language models, yielding state-of-the-art results.
今天的网络用户每天都会接触到误导和宣传的新闻文章和媒体帖子。为了应对这种情况,已经设计了一些方法,旨在实现更健康、更安全的在线新闻和媒体消费。自动系统能够支持人类检测此类内容;然而,广泛采用这些系统的一个主要障碍是,除了准确之外,这些系统的决定还需要是可解释的,以便得到用户的信任和广泛采用。由于误导和宣传内容通过使用一些欺骗技术来影响读者,我们建议检测并展示这些技术的使用,以提供可解释性。特别是,我们定义了定性描述特征,并分析了它们对检测欺骗技术的适用性。我们进一步表明,我们的可解释特征可以很容易地与预训练的语言模型相结合,从而产生最先进的结果。
{"title":"Interpretable Propaganda Detection in News Articles","authors":"Seunghak Yu, Giovanni Da San Martino, Mitra Mohtarami, James R. Glass, Preslav Nakov","doi":"10.26615/978-954-452-072-4_179","DOIUrl":"https://doi.org/10.26615/978-954-452-072-4_179","url":null,"abstract":"Online users today are exposed to misleading and propagandistic news articles and media posts on a daily basis. To counter thus, a number of approaches have been designed aiming to achieve a healthier and safer online news and media consumption. Automatic systems are able to support humans in detecting such content; yet, a major impediment to their broad adoption is that besides being accurate, the decisions of such systems need also to be interpretable in order to be trusted and widely adopted by users. Since misleading and propagandistic content influences readers through the use of a number of deception techniques, we propose to detect and to show the use of such techniques as a way to offer interpretability. In particular, we define qualitatively descriptive features and we analyze their suitability for detecting deception techniques. We further show that our interpretable features can be easily combined with pre-trained language models, yielding state-of-the-art results.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126193013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Can the Transformer Be Used as a Drop-in Replacement for RNNs in Text-Generating GANs? 转换器可以作为文本生成gan中rnn的替代吗?
Pub Date : 2021-08-26 DOI: 10.26615/978-954-452-072-4_021
Kevin Blin, Andrei Kucharavy
In this paper we address the problem of fine-tuned text generation with a limited computational budget. For that, we use a well-performing text generative adversarial network (GAN) architecture - Diversity-Promoting GAN (DPGAN), and attempted a drop-in replacement of the LSTM layer with a self-attention-based Transformer layer in order to leverage their efficiency. The resulting Self-Attention DPGAN (SADPGAN) was evaluated for performance, quality and diversity of generated text and stability. Computational experiments suggested that a transformer architecture is unable to drop-in replace the LSTM layer, under-performing during the pre-training phase and undergoing a complete mode collapse during the GAN tuning phase. Our results suggest that the transformer architecture need to be adapted before it can be used as a replacement for RNNs in text-generating GANs.
在本文中,我们解决了在有限的计算预算下微调文本生成的问题。为此,我们使用了一种性能良好的文本生成对抗网络(GAN)架构——Diversity-Promoting GAN (DPGAN),并尝试用基于自注意力的Transformer层替代LSTM层,以利用它们的效率。对生成的自注意DPGAN (SADPGAN)的性能、质量和多样性以及稳定性进行了评估。计算实验表明,变压器结构不能直接取代LSTM层,在预训练阶段表现不佳,在GAN调谐阶段经历完全的模式崩溃。我们的研究结果表明,在文本生成gan中用作rnn的替代品之前,需要对变压器架构进行调整。
{"title":"Can the Transformer Be Used as a Drop-in Replacement for RNNs in Text-Generating GANs?","authors":"Kevin Blin, Andrei Kucharavy","doi":"10.26615/978-954-452-072-4_021","DOIUrl":"https://doi.org/10.26615/978-954-452-072-4_021","url":null,"abstract":"In this paper we address the problem of fine-tuned text generation with a limited computational budget. For that, we use a well-performing text generative adversarial network (GAN) architecture - Diversity-Promoting GAN (DPGAN), and attempted a drop-in replacement of the LSTM layer with a self-attention-based Transformer layer in order to leverage their efficiency. The resulting Self-Attention DPGAN (SADPGAN) was evaluated for performance, quality and diversity of generated text and stability. Computational experiments suggested that a transformer architecture is unable to drop-in replace the LSTM layer, under-performing during the pre-training phase and undergoing a complete mode collapse during the GAN tuning phase. Our results suggest that the transformer architecture need to be adapted before it can be used as a replacement for RNNs in text-generating GANs.","PeriodicalId":284493,"journal":{"name":"Recent Advances in Natural Language Processing","volume":"81 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131333959","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Recent Advances in Natural Language Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1