首页 > 最新文献

Proceedings of the The First Workshop on Ever Evolving NLP (EvoNLP)最新文献

英文 中文
MLLabs-LIG at TempoWiC 2022: A Generative Approach for Examining Temporal Meaning Shift MLLabs-LIG在TempoWiC 2022:一种用于检查时间意义转移的生成方法
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.evonlp-1.1
Chenyang Lyu, Yongxin Zhou, Tianbo Ji
In this paper, we present our system for the EvoNLP 2022 shared task Temporal Meaning Shift (TempoWiC). Different from the typically used discriminative model, we propose a generative approach based on pre-trained generation models. The basic architecture of our system is a seq2seq model where the input sequence consists of two documents followed by a question asking whether the meaning of target word changed or not, the target output sequence is a declarative sentence describing the meaning of target word changed or not. The experimental results on TempoWiC test set show that our best system (with time information) obtained an accuracy and Marco F-1 score of 68.09% and 62.59% respectively, which ranked 12th among all submitted systems. The results have shown the plausibility of using generation model for WiC tasks, meanwhile also indicate there’s still room for further improvement.
在本文中,我们提出了EvoNLP 2022共享任务时间意义转换(TempoWiC)的系统。与通常使用的判别模型不同,我们提出了一种基于预训练生成模型的生成方法。我们的系统的基本架构是一个seq2seq模型,其中输入序列由两个文档组成,后面跟着一个询问目标单词的含义是否改变的问题,目标输出序列是一个描述目标单词的含义是否改变的陈述句。TempoWiC测试集上的实验结果表明,我们的最佳系统(含时间信息)的准确率和Marco F-1分数分别为68.09%和62.59%,在所有提交的系统中排名第12位。研究结果表明了生成模型在WiC任务中的可行性,同时也表明了该模型仍有进一步改进的空间。
{"title":"MLLabs-LIG at TempoWiC 2022: A Generative Approach for Examining Temporal Meaning Shift","authors":"Chenyang Lyu, Yongxin Zhou, Tianbo Ji","doi":"10.18653/v1/2022.evonlp-1.1","DOIUrl":"https://doi.org/10.18653/v1/2022.evonlp-1.1","url":null,"abstract":"In this paper, we present our system for the EvoNLP 2022 shared task Temporal Meaning Shift (TempoWiC). Different from the typically used discriminative model, we propose a generative approach based on pre-trained generation models. The basic architecture of our system is a seq2seq model where the input sequence consists of two documents followed by a question asking whether the meaning of target word changed or not, the target output sequence is a declarative sentence describing the meaning of target word changed or not. The experimental results on TempoWiC test set show that our best system (with time information) obtained an accuracy and Marco F-1 score of 68.09% and 62.59% respectively, which ranked 12th among all submitted systems. The results have shown the plausibility of using generation model for WiC tasks, meanwhile also indicate there’s still room for further improvement.","PeriodicalId":158578,"journal":{"name":"Proceedings of the The First Workshop on Ever Evolving NLP (EvoNLP)","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114871985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
CC-Top: Constrained Clustering for Dynamic Topic Discovery CC-Top:动态主题发现的约束聚类
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.evonlp-1.5
Jann Goschenhofer, Pranav Ragupathy, C. Heumann, Bernd Bischl, M. Aßenmacher
Research on multi-class text classification of short texts mainly focuses on supervised (transfer) learning approaches, requiring a finite set of pre-defined classes which is constant over time. This work explores deep constrained clustering (CC) as an alternative to supervised learning approaches in a setting with a dynamically changing number of classes, a task we introduce as dynamic topic discovery (DTD).We do so by using pairwise similarity constraints instead of instance-level class labels which allow for a flexible number of classes while exhibiting a competitive performance compared to supervised approaches. First, we substantiate this through a series of experiments and show that CC algorithms exhibit a predictive performance similar to state-of-the-art supervised learning algorithms while requiring less annotation effort.Second, we demonstrate the overclustering capabilities of deep CC for detecting topics in short text data sets in the absence of the ground truth class cardinality during model training.Third, we showcase that these capabilities can be leveraged for the DTD setting as a step towards dynamic learning over time and finally, we release our codebase to nurture further research in this area.
短文本多类文本分类的研究主要集中在监督(迁移)学习方法上,需要有限的预定义类集,这些类集随着时间的推移是恒定的。这项工作探索了深度约束聚类(CC)作为监督学习方法的替代方法,在类数量动态变化的环境中,我们将这种任务称为动态主题发现(DTD)。我们通过使用两两相似约束而不是实例级类标签来实现这一点,这允许灵活的类数量,同时与监督方法相比表现出具有竞争力的性能。首先,我们通过一系列实验证实了这一点,并表明CC算法表现出与最先进的监督学习算法相似的预测性能,同时需要更少的注释工作。其次,我们展示了深度CC在模型训练期间缺乏基础真类基数的情况下在短文本数据集中检测主题的过度聚类能力。第三,我们展示了这些功能可以用于DTD设置,作为实现动态学习的一个步骤,最后,我们发布了我们的代码库,以促进该领域的进一步研究。
{"title":"CC-Top: Constrained Clustering for Dynamic Topic Discovery","authors":"Jann Goschenhofer, Pranav Ragupathy, C. Heumann, Bernd Bischl, M. Aßenmacher","doi":"10.18653/v1/2022.evonlp-1.5","DOIUrl":"https://doi.org/10.18653/v1/2022.evonlp-1.5","url":null,"abstract":"Research on multi-class text classification of short texts mainly focuses on supervised (transfer) learning approaches, requiring a finite set of pre-defined classes which is constant over time. This work explores deep constrained clustering (CC) as an alternative to supervised learning approaches in a setting with a dynamically changing number of classes, a task we introduce as dynamic topic discovery (DTD).We do so by using pairwise similarity constraints instead of instance-level class labels which allow for a flexible number of classes while exhibiting a competitive performance compared to supervised approaches. First, we substantiate this through a series of experiments and show that CC algorithms exhibit a predictive performance similar to state-of-the-art supervised learning algorithms while requiring less annotation effort.Second, we demonstrate the overclustering capabilities of deep CC for detecting topics in short text data sets in the absence of the ground truth class cardinality during model training.Third, we showcase that these capabilities can be leveraged for the DTD setting as a step towards dynamic learning over time and finally, we release our codebase to nurture further research in this area.","PeriodicalId":158578,"journal":{"name":"Proceedings of the The First Workshop on Ever Evolving NLP (EvoNLP)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129478844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging time-dependent lexical features for offensive language detection 利用时间相关的词汇特征进行攻击性语言检测
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.evonlp-1.7
Barbara McGillivray, Malithi Alahapperuma, Jonathan Cook, C. Di Bonaventura, Albert Meroño-Peñuela, Gareth Tyson, Steven R. Wilson
We present a study on the integration of time-sensitive information in lexicon-based offensive language detection systems. Our focus is on Offenseval sub-task A, aimed at detecting offensive tweets. We apply a semantic change detection algorithm over a short time span of two years to detect words whose semantics has changed and we focus particularly on those words that acquired or lost an offensive meaning between 2019 and 2020. Using the output of this semantic change detection approach, we train an SVM classifier on the Offenseval 2019 training set. We build on the already competitive SINAI system submitted to Offenseval 2019 by adding new lexical features, including those that capture the change in usage of words and their association with emerging offensive usages. We discuss the challenges, opportunities and limitations of integrating semantic change detection in offensive language detection models. Our work draws attention to an often neglected aspect of offensive language, namely that the meanings of words are constantly evolving and that NLP systems that account for this change can achieve good performance even when not trained on the most recent training data.
本文对基于词典的攻击性语言检测系统中时间敏感信息的集成进行了研究。我们的重点是攻击性子任务A,旨在检测攻击性推文。我们在两年的短时间内应用语义变化检测算法来检测语义发生变化的单词,我们特别关注那些在2019年至2020年期间获得或失去冒犯性含义的单词。使用这种语义变化检测方法的输出,我们在Offenseval 2019训练集上训练SVM分类器。我们在提交给Offenseval 2019的竞争性西奈系统的基础上,增加了新的词汇功能,包括那些捕捉单词用法变化及其与新出现的攻击性用法的关联的功能。我们讨论了在攻击性语言检测模型中集成语义变化检测的挑战、机遇和局限性。我们的工作引起了人们对攻击性语言一个经常被忽视的方面的关注,即单词的含义是不断变化的,而考虑到这种变化的NLP系统即使在没有接受最新训练数据训练的情况下也能取得良好的表现。
{"title":"Leveraging time-dependent lexical features for offensive language detection","authors":"Barbara McGillivray, Malithi Alahapperuma, Jonathan Cook, C. Di Bonaventura, Albert Meroño-Peñuela, Gareth Tyson, Steven R. Wilson","doi":"10.18653/v1/2022.evonlp-1.7","DOIUrl":"https://doi.org/10.18653/v1/2022.evonlp-1.7","url":null,"abstract":"We present a study on the integration of time-sensitive information in lexicon-based offensive language detection systems. Our focus is on Offenseval sub-task A, aimed at detecting offensive tweets. We apply a semantic change detection algorithm over a short time span of two years to detect words whose semantics has changed and we focus particularly on those words that acquired or lost an offensive meaning between 2019 and 2020. Using the output of this semantic change detection approach, we train an SVM classifier on the Offenseval 2019 training set. We build on the already competitive SINAI system submitted to Offenseval 2019 by adding new lexical features, including those that capture the change in usage of words and their association with emerging offensive usages. We discuss the challenges, opportunities and limitations of integrating semantic change detection in offensive language detection models. Our work draws attention to an often neglected aspect of offensive language, namely that the meanings of words are constantly evolving and that NLP systems that account for this change can achieve good performance even when not trained on the most recent training data.","PeriodicalId":158578,"journal":{"name":"Proceedings of the The First Workshop on Ever Evolving NLP (EvoNLP)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114671732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Class Incremental Learning for Intent Classification with Limited or No Old Data 利用有限或无旧数据进行意图分类的类递增学习
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.evonlp-1.4
Debjit Paul, Daniil Sorokin, Judith Gaspers
In this paper, we explore class-incremental learning for intent classification (IC) in a setting with limited old data available. IC is the task of mapping user utterances to their corresponding intents. Even though class-incremental learning without storing the old data yields high potential of reducing human and computational resources in industry NLP model releases, to the best of our knowledge, it hasn’t been studied for NLP classification tasks in the literature before. In this work, we compare several contemporary class-incremental learning methods, i.e., BERT warm start, L2, Elastic Weight Consolidation, RecAdam and Knowledge Distillation within two realistic class-incremental learning scenarios: one where only the previous model is assumed to be available, but no data corresponding to old classes, and one in which limited unlabeled data for old classes is assumed to be available. Our results indicate that among the investigated continual learning methods, Knowledge Distillation worked best for our class-incremental learning tasks, and adding limited unlabeled data helps the model in both adaptability and stability.
在本文中,我们探讨了在旧数据有限的情况下进行意图分类(IC)的类递增学习。意图分类是将用户话语映射到其相应意图的任务。尽管不存储旧数据的类递增学习在减少行业 NLP 模型发布的人力和计算资源方面具有很大的潜力,但据我们所知,以前的文献中还没有针对 NLP 分类任务进行过研究。在这项工作中,我们在两个现实的类增量学习场景中比较了几种当代的类增量学习方法,即 BERT warm start、L2、Elastic Weight Consolidation、RecAdam 和 Knowledge Distillation:一个场景是假定只有之前的模型可用,但没有与旧类相对应的数据,另一个场景是假定有有限的未标记旧类数据可用。我们的结果表明,在所研究的持续学习方法中,知识蒸馏法最适合我们的类递增学习任务,而且添加有限的未标记数据有助于模型的适应性和稳定性。
{"title":"Class Incremental Learning for Intent Classification with Limited or No Old Data","authors":"Debjit Paul, Daniil Sorokin, Judith Gaspers","doi":"10.18653/v1/2022.evonlp-1.4","DOIUrl":"https://doi.org/10.18653/v1/2022.evonlp-1.4","url":null,"abstract":"In this paper, we explore class-incremental learning for intent classification (IC) in a setting with limited old data available. IC is the task of mapping user utterances to their corresponding intents. Even though class-incremental learning without storing the old data yields high potential of reducing human and computational resources in industry NLP model releases, to the best of our knowledge, it hasn’t been studied for NLP classification tasks in the literature before. In this work, we compare several contemporary class-incremental learning methods, i.e., BERT warm start, L2, Elastic Weight Consolidation, RecAdam and Knowledge Distillation within two realistic class-incremental learning scenarios: one where only the previous model is assumed to be available, but no data corresponding to old classes, and one in which limited unlabeled data for old classes is assumed to be available. Our results indicate that among the investigated continual learning methods, Knowledge Distillation worked best for our class-incremental learning tasks, and adding limited unlabeled data helps the model in both adaptability and stability.","PeriodicalId":158578,"journal":{"name":"Proceedings of the The First Workshop on Ever Evolving NLP (EvoNLP)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126306542","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HSE at TempoWiC: Detecting Meaning Shift in Social Media with Diachronic Language Models HSE在TempoWiC:用历时语言模型检测社交媒体的意义转移
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.evonlp-1.6
Elizaveta Tukhtina, Kseniia Kashleva, Svetlana Vydrina
This paper describes our methods for temporal meaning shift detection, implemented during the TempoWiC shared task. We present two systems: with and without time span data usage. Our approaches are based on the language models fine-tuned for Twitter domain. Both systems outperformed all the competition’s baselines except TimeLMs-SIM. Our best submission achieved the macro-F1 score of 70.09% and took the 7th place. This result was achieved by using diachronic language models from the TimeLMs project.
本文描述了我们在tempoic共享任务中实现的时间意义偏移检测方法。我们提出了两个系统:有和没有时间跨度的数据使用。我们的方法基于针对Twitter领域进行微调的语言模型。除了TimeLMs-SIM之外,这两个系统都超过了所有竞争对手的基准。我们最好的投稿获得了70.09%的macro-F1分数,获得了第7名。这个结果是通过使用来自TimeLMs项目的历时语言模型实现的。
{"title":"HSE at TempoWiC: Detecting Meaning Shift in Social Media with Diachronic Language Models","authors":"Elizaveta Tukhtina, Kseniia Kashleva, Svetlana Vydrina","doi":"10.18653/v1/2022.evonlp-1.6","DOIUrl":"https://doi.org/10.18653/v1/2022.evonlp-1.6","url":null,"abstract":"This paper describes our methods for temporal meaning shift detection, implemented during the TempoWiC shared task. We present two systems: with and without time span data usage. Our approaches are based on the language models fine-tuned for Twitter domain. Both systems outperformed all the competition’s baselines except TimeLMs-SIM. Our best submission achieved the macro-F1 score of 70.09% and took the 7th place. This result was achieved by using diachronic language models from the TimeLMs project.","PeriodicalId":158578,"journal":{"name":"Proceedings of the The First Workshop on Ever Evolving NLP (EvoNLP)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127327656","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the The First Workshop on Ever Evolving NLP (EvoNLP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1