首页 > 最新文献

Proceedings of COLING. International Conference on Computational Linguistics最新文献

英文 中文
Natural Language Inference Prompts for Zero-shot Emotion Classification in Text across Corpora 跨语料库文本零概率情感分类的自然语言推理提示
Pub Date : 2022-09-14 DOI: 10.48550/arXiv.2209.06701
F. Plaza-Del-Arco, Mar'ia-Teresa Mart'in-Valdivia, Roman Klinger
Within textual emotion classification, the set of relevant labels depends on the domain and application scenario and might not be known at the time of model development. This conflicts with the classical paradigm of supervised learning in which the labels need to be predefined. A solution to obtain a model with a flexible set of labels is to use the paradigm of zero-shot learning as a natural language inference task, which in addition adds the advantage of not needing any labeled training data. This raises the question how to prompt a natural language inference model for zero-shot learning emotion classification. Options for prompt formulations include the emotion name anger alone or the statement “This text expresses anger”. With this paper, we analyze how sensitive a natural language inference-based zero-shot-learning classifier is to such changes to the prompt under consideration of the corpus: How carefully does the prompt need to be selected? We perform experiments on an established set of emotion datasets presenting different language registers according to different sources (tweets, events, blogs) with three natural language inference models and show that indeed the choice of a particular prompt formulation needs to fit to the corpus. We show that this challenge can be tackled with combinations of multiple prompts. Such ensemble is more robust across corpora than individual prompts and shows nearly the same performance as the individual best prompt for a particular corpus.
在文本情感分类中,相关标签的集合取决于领域和应用场景,在模型开发时可能不知道。这与监督学习的经典范例相冲突,在这种范例中,标签需要预先定义。获得具有灵活标签集的模型的一种解决方案是使用零次学习范式作为自然语言推理任务,这还增加了不需要任何标记训练数据的优点。这就提出了一个问题,即如何为零学习情绪分类提示一个自然语言推理模型。提示公式的选项包括单独的情绪名称愤怒或声明“这篇文章表达愤怒”。在本文中,我们分析了基于自然语言推理的零学习分类器在考虑语料库的情况下,对提示的这种变化有多敏感:需要多仔细地选择提示?我们在一组已建立的情感数据集上进行了实验,这些数据集根据不同的来源(推文、事件、博客)使用三种自然语言推理模型呈现不同的语言寄存器,并表明特定提示公式的选择确实需要适合语料库。我们展示了这个挑战可以通过多个提示的组合来解决。这种集成在整个语料库中比单个提示更健壮,并且显示出与特定语料库的单个最佳提示几乎相同的性能。
{"title":"Natural Language Inference Prompts for Zero-shot Emotion Classification in Text across Corpora","authors":"F. Plaza-Del-Arco, Mar'ia-Teresa Mart'in-Valdivia, Roman Klinger","doi":"10.48550/arXiv.2209.06701","DOIUrl":"https://doi.org/10.48550/arXiv.2209.06701","url":null,"abstract":"Within textual emotion classification, the set of relevant labels depends on the domain and application scenario and might not be known at the time of model development. This conflicts with the classical paradigm of supervised learning in which the labels need to be predefined. A solution to obtain a model with a flexible set of labels is to use the paradigm of zero-shot learning as a natural language inference task, which in addition adds the advantage of not needing any labeled training data. This raises the question how to prompt a natural language inference model for zero-shot learning emotion classification. Options for prompt formulations include the emotion name anger alone or the statement “This text expresses anger”. With this paper, we analyze how sensitive a natural language inference-based zero-shot-learning classifier is to such changes to the prompt under consideration of the corpus: How carefully does the prompt need to be selected? We perform experiments on an established set of emotion datasets presenting different language registers according to different sources (tweets, events, blogs) with three natural language inference models and show that indeed the choice of a particular prompt formulation needs to fit to the corpus. We show that this challenge can be tackled with combinations of multiple prompts. Such ensemble is more robust across corpora than individual prompts and shows nearly the same performance as the individual best prompt for a particular corpus.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"14 1","pages":"6805-6817"},"PeriodicalIF":0.0,"publicationDate":"2022-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72870414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
COMMA: Modeling Relationship among Motivations, Emotions and Actions in Language-based Human Activities 逗号:基于语言的人类活动中动机、情绪和行为之间的关系建模
Pub Date : 2022-09-14 DOI: 10.48550/arXiv.2209.06470
Yuqiang Xie, Yue Hu, Wei Peng, Guanqun Bi, Luxi Xing
Motivations, emotions, and actions are inter-related essential factors in human activities. While motivations and emotions have long been considered at the core of exploring how people take actions in human activities, there has been relatively little research supporting analyzing the relationship between human mental states and actions. We present the first study that investigates the viability of modeling motivations, emotions, and actions in language-based human activities, named COMMA (Cognitive Framework of Human Activities). Guided by COMMA, we define three natural language processing tasks (emotion understanding, motivation understanding and conditioned action generation), and build a challenging dataset Hail through automatically extracting samples from Story Commonsense. Experimental results on NLP applications prove the effectiveness of modeling the relationship. Furthermore, our models inspired by COMMA can better reveal the essential relationship among motivations, emotions and actions than existing methods.
动机、情绪和行动是人类活动中相互关联的基本因素。虽然动机和情绪一直被认为是探索人们在人类活动中如何采取行动的核心,但支持分析人类心理状态和行动之间关系的研究相对较少。我们提出了第一项研究,调查了在基于语言的人类活动中建模动机、情感和行为的可行性,名为COMMA(人类活动认知框架)。在逗号的引导下,我们定义了三个自然语言处理任务(情感理解、动机理解和条件动作生成),并通过从故事常识中自动提取样本来构建具有挑战性的数据集Hail。在自然语言处理应用中的实验结果证明了这种关系建模的有效性。此外,受逗号启发的模型比现有方法更能揭示动机、情绪和行为之间的本质关系。
{"title":"COMMA: Modeling Relationship among Motivations, Emotions and Actions in Language-based Human Activities","authors":"Yuqiang Xie, Yue Hu, Wei Peng, Guanqun Bi, Luxi Xing","doi":"10.48550/arXiv.2209.06470","DOIUrl":"https://doi.org/10.48550/arXiv.2209.06470","url":null,"abstract":"Motivations, emotions, and actions are inter-related essential factors in human activities. While motivations and emotions have long been considered at the core of exploring how people take actions in human activities, there has been relatively little research supporting analyzing the relationship between human mental states and actions. We present the first study that investigates the viability of modeling motivations, emotions, and actions in language-based human activities, named COMMA (Cognitive Framework of Human Activities). Guided by COMMA, we define three natural language processing tasks (emotion understanding, motivation understanding and conditioned action generation), and build a challenging dataset Hail through automatically extracting samples from Story Commonsense. Experimental results on NLP applications prove the effectiveness of modeling the relationship. Furthermore, our models inspired by COMMA can better reveal the essential relationship among motivations, emotions and actions than existing methods.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"556 1","pages":"163-177"},"PeriodicalIF":0.0,"publicationDate":"2022-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77289969","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Prompt Combines Paraphrase: Teaching Pre-trained Models to Understand Rare Biomedical Words 提示结合释义:教授预先训练的模型来理解罕见的生物医学词汇
Pub Date : 2022-09-14 DOI: 10.48550/arXiv.2209.06453
Hao Wang, Chi-Liang Liu, Nuwa Xi, Sendong Zhao, Meizhi Ju, Shiwei Zhang, Ziheng Zhang, Yefeng Zheng, Bing Qin, Ting Liu
Prompt-based fine-tuning for pre-trained models has proven effective for many natural language processing tasks under few-shot settings in general domain. However, tuning with prompt in biomedical domain has not been investigated thoroughly. Biomedical words are often rare in general domain, but quite ubiquitous in biomedical contexts, which dramatically deteriorates the performance of pre-trained models on downstream biomedical applications even after fine-tuning, especially in low-resource scenarios. We propose a simple yet effective approach to helping models learn rare biomedical words during tuning with prompt. Experimental results show that our method can achieve up to 6% improvement in biomedical natural language inference task without any extra parameters or training steps using few-shot vanilla prompt settings.
预训练模型的基于提示的微调已被证明对一般领域中许多自然语言处理任务在少数镜头设置下是有效的。然而,在生物医学领域,对提示调谐的研究还不够深入。生物医学词汇通常在一般领域很少出现,但在生物医学语境中却非常普遍,这大大降低了预训练模型在下游生物医学应用中的性能,即使经过微调,特别是在低资源场景下。我们提出了一种简单而有效的方法来帮助模型在调整过程中学习罕见的生物医学词汇。实验结果表明,该方法在不需要任何额外参数或训练步骤的情况下,使用少量的提示设置,可以将生物医学自然语言推理任务提高6%。
{"title":"Prompt Combines Paraphrase: Teaching Pre-trained Models to Understand Rare Biomedical Words","authors":"Hao Wang, Chi-Liang Liu, Nuwa Xi, Sendong Zhao, Meizhi Ju, Shiwei Zhang, Ziheng Zhang, Yefeng Zheng, Bing Qin, Ting Liu","doi":"10.48550/arXiv.2209.06453","DOIUrl":"https://doi.org/10.48550/arXiv.2209.06453","url":null,"abstract":"Prompt-based fine-tuning for pre-trained models has proven effective for many natural language processing tasks under few-shot settings in general domain. However, tuning with prompt in biomedical domain has not been investigated thoroughly. Biomedical words are often rare in general domain, but quite ubiquitous in biomedical contexts, which dramatically deteriorates the performance of pre-trained models on downstream biomedical applications even after fine-tuning, especially in low-resource scenarios. We propose a simple yet effective approach to helping models learn rare biomedical words during tuning with prompt. Experimental results show that our method can achieve up to 6% improvement in biomedical natural language inference task without any extra parameters or training steps using few-shot vanilla prompt settings.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"209 1","pages":"1422-1431"},"PeriodicalIF":0.0,"publicationDate":"2022-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86843467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Generalized Intent Discovery: Learning from Open World Dialogue System 广义意图发现:从开放世界对话系统学习
Pub Date : 2022-09-13 DOI: 10.48550/arXiv.2209.06030
Yutao Mou, Keqing He, Yanan Wu, Pei Wang, Jingang Wang, Wei Wu, Y. Huang, Junlan Feng, Weiran Xu
Traditional intent classification models are based on a pre-defined intent set and only recognize limited in-domain (IND) intent classes. But users may input out-of-domain (OOD) queries in a practical dialogue system. Such OOD queries can provide directions for future improvement. In this paper, we define a new task, Generalized Intent Discovery (GID), which aims to extend an IND intent classifier to an open-world intent set including IND and OOD intents. We hope to simultaneously classify a set of labeled IND intent classes while discovering and recognizing new unlabeled OOD types incrementally. We construct three public datasets for different application scenarios and propose two kinds of frameworks, pipeline-based and end-to-end for future work. Further, we conduct exhaustive experiments and qualitative analysis to comprehend key challenges and provide new guidance for future GID research.
传统的意图分类模型基于预定义的意图集,只能识别有限的域内(IND)意图类。但是用户可以在实际的对话系统中输入域外(OOD)查询。这样的OOD查询可以为未来的改进提供方向。在本文中,我们定义了一个新的任务,即广义意图发现(GID),它旨在将IND意图分类器扩展到包含IND和OOD意图的开放世界意图集。我们希望在发现和识别新的未标记的OOD类型的同时,对一组标记的IND意图类进行分类。我们针对不同的应用场景构建了三个公共数据集,并为未来的工作提出了基于管道和端到端两种框架。此外,我们进行详尽的实验和定性分析,以了解关键挑战,并为未来的GID研究提供新的指导。
{"title":"Generalized Intent Discovery: Learning from Open World Dialogue System","authors":"Yutao Mou, Keqing He, Yanan Wu, Pei Wang, Jingang Wang, Wei Wu, Y. Huang, Junlan Feng, Weiran Xu","doi":"10.48550/arXiv.2209.06030","DOIUrl":"https://doi.org/10.48550/arXiv.2209.06030","url":null,"abstract":"Traditional intent classification models are based on a pre-defined intent set and only recognize limited in-domain (IND) intent classes. But users may input out-of-domain (OOD) queries in a practical dialogue system. Such OOD queries can provide directions for future improvement. In this paper, we define a new task, Generalized Intent Discovery (GID), which aims to extend an IND intent classifier to an open-world intent set including IND and OOD intents. We hope to simultaneously classify a set of labeled IND intent classes while discovering and recognizing new unlabeled OOD types incrementally. We construct three public datasets for different application scenarios and propose two kinds of frameworks, pipeline-based and end-to-end for future work. Further, we conduct exhaustive experiments and qualitative analysis to comprehend key challenges and provide new guidance for future GID research.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"50 1","pages":"707-720"},"PeriodicalIF":0.0,"publicationDate":"2022-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84262993","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
CSL: A Large-scale Chinese Scientific Literature Dataset CSL:大型中文科学文献数据集
Pub Date : 2022-09-12 DOI: 10.48550/arXiv.2209.05034
Yudong Li, Yuqing Zhang, Zhe Zhao, Lin-cheng Shen, Weijie Liu, Weiquan Mao, Hui Zhang
Scientific literature serves as a high-quality corpus, supporting a lot of Natural Language Processing (NLP) research. However, existing datasets are centered around the English language, which restricts the development of Chinese scientific NLP. In this work, we present CSL, a large-scale Chinese Scientific Literature dataset, which contains the titles, abstracts, keywords and academic fields of 396k papers. To our knowledge, CSL is the first scientific document dataset in Chinese. The CSL can serve as a Chinese corpus. Also, this semi-structured data is a natural annotation that can constitute many supervised NLP tasks. Based on CSL, we present a benchmark to evaluate the performance of models across scientific domain tasks, i.e., summarization, keyword generation and text classification. We analyze the behavior of existing text-to-text models on the evaluation tasks and reveal the challenges for Chinese scientific NLP tasks, which provides a valuable reference for future research. Data and code will be publicly available.
科学文献是一个高质量的语料库,支持了许多自然语言处理(NLP)研究。然而,现有的数据集以英语为中心,制约了中国科学自然语言处理的发展。在本工作中,我们建立了一个大型中文科学文献数据集CSL,该数据集包含39.6万篇论文的标题、摘要、关键词和学术领域。据我们所知,CSL是第一个中文科学文献数据集。该语言库可以作为汉语语料库。此外,这种半结构化数据是一种自然注释,可以构成许多有监督的NLP任务。基于CSL,我们提出了一个评估模型跨科学领域任务(即摘要、关键字生成和文本分类)性能的基准。分析了现有文本到文本模型在评价任务上的行为,揭示了中国科学NLP任务面临的挑战,为未来的研究提供了有价值的参考。数据和代码将公开提供。
{"title":"CSL: A Large-scale Chinese Scientific Literature Dataset","authors":"Yudong Li, Yuqing Zhang, Zhe Zhao, Lin-cheng Shen, Weijie Liu, Weiquan Mao, Hui Zhang","doi":"10.48550/arXiv.2209.05034","DOIUrl":"https://doi.org/10.48550/arXiv.2209.05034","url":null,"abstract":"Scientific literature serves as a high-quality corpus, supporting a lot of Natural Language Processing (NLP) research. However, existing datasets are centered around the English language, which restricts the development of Chinese scientific NLP. In this work, we present CSL, a large-scale Chinese Scientific Literature dataset, which contains the titles, abstracts, keywords and academic fields of 396k papers. To our knowledge, CSL is the first scientific document dataset in Chinese. The CSL can serve as a Chinese corpus. Also, this semi-structured data is a natural annotation that can constitute many supervised NLP tasks. Based on CSL, we present a benchmark to evaluate the performance of models across scientific domain tasks, i.e., summarization, keyword generation and text classification. We analyze the behavior of existing text-to-text models on the evaluation tasks and reveal the challenges for Chinese scientific NLP tasks, which provides a valuable reference for future research. Data and code will be publicly available.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"7 1","pages":"3917-3923"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86438716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
SANCL: Multimodal Review Helpfulness Prediction with Selective Attention and Natural Contrastive Learning 选择性注意与自然对比学习的多模态复习帮助性预测
Pub Date : 2022-09-12 DOI: 10.48550/arXiv.2209.05040
Wei Han, Hui Chen, Zhen Hai, Soujanya Poria, Lidong Bing
With the boom of e-commerce, Multimodal Review Helpfulness Prediction (MRHP) that identifies the helpfulness score of multimodal product reviews has become a research hotspot. Previous work on this task focuses on attention-based modality fusion, information integration, and relation modeling, which primarily exposes the following drawbacks: 1) the model may fail to capture the really essential information due to its indiscriminate attention formulation; 2) lack appropriate modeling methods that takes full advantage of correlation among provided data. In this paper, we propose SANCL: Selective Attention and Natural Contrastive Learning for MRHP. SANCL adopts a probe-based strategy to enforce high attention weights on the regions of greater significance. It also constructs a contrastive learning framework based on natural matching properties in the dataset. Experimental results on two benchmark datasets with three categories show that SANCL achieves state-of-the-art baseline performance with lower memory consumption.
随着电子商务的蓬勃发展,识别多模式产品评论有用性分数的多模式评论有用性预测(MRHP)成为研究热点。以往的研究主要集中在基于注意的模态融合、信息集成和关系建模等方面,主要暴露出以下缺陷:1)模型由于不加区分的注意表述,可能无法捕捉到真正重要的信息;2)缺乏适当的建模方法,不能充分利用所提供数据之间的相关性。在本文中,我们提出了sanl:选择性注意和自然对比学习。SANCL采用基于探针的策略,在更重要的区域上强制执行高关注权。并基于数据集的自然匹配属性构建了一个对比学习框架。在两个包含三个类别的基准数据集上的实验结果表明,SANCL以较低的内存消耗达到了最先进的基准性能。
{"title":"SANCL: Multimodal Review Helpfulness Prediction with Selective Attention and Natural Contrastive Learning","authors":"Wei Han, Hui Chen, Zhen Hai, Soujanya Poria, Lidong Bing","doi":"10.48550/arXiv.2209.05040","DOIUrl":"https://doi.org/10.48550/arXiv.2209.05040","url":null,"abstract":"With the boom of e-commerce, Multimodal Review Helpfulness Prediction (MRHP) that identifies the helpfulness score of multimodal product reviews has become a research hotspot. Previous work on this task focuses on attention-based modality fusion, information integration, and relation modeling, which primarily exposes the following drawbacks: 1) the model may fail to capture the really essential information due to its indiscriminate attention formulation; 2) lack appropriate modeling methods that takes full advantage of correlation among provided data. In this paper, we propose SANCL: Selective Attention and Natural Contrastive Learning for MRHP. SANCL adopts a probe-based strategy to enforce high attention weights on the regions of greater significance. It also constructs a contrastive learning framework based on natural matching properties in the dataset. Experimental results on two benchmark datasets with three categories show that SANCL achieves state-of-the-art baseline performance with lower memory consumption.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"7 1","pages":"5666-5677"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87688584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Adaptive Meta-learner via Gradient Similarity for Few-shot Text Classification 基于梯度相似度的小样本文本分类自适应元学习器
Pub Date : 2022-09-10 DOI: 10.48550/arXiv.2209.04702
Tianyi Lei, Honghui Hu, Qiaoyang Luo, Dezhong Peng, Xu Wang
Few-shot text classification aims to classify the text under the few-shot scenario. Most of the previous methods adopt optimization-based meta learning to obtain task distribution. However, due to the neglect of matching between the few amount of samples and complicated models, as well as the distinction between useful and useless task features, these methods suffer from the overfitting issue. To address this issue, we propose a novel Adaptive Meta-learner via Gradient Similarity (AMGS) method to improve the model generalization ability to a new task. Specifically, the proposed AMGS alleviates the overfitting based on two aspects: (i) acquiring the potential semantic representation of samples and improving model generalization through the self-supervised auxiliary task in the inner loop, (ii) leveraging the adaptive meta-learner via gradient similarity to add constraints on the gradient obtained by base-learner in the outer loop. Moreover, we make a systematic analysis of the influence of regularization on the entire framework. Experimental results on several benchmarks demonstrate that the proposed AMGS consistently improves few-shot text classification performance compared with the state-of-the-art optimization-based meta-learning approaches. The code is available at: https://github.com/Tianyi-Lei.
少镜头文本分类的目的是对少镜头场景下的文本进行分类。以往的方法大多采用基于优化的元学习来获得任务分布。然而,由于忽略了少量样本和复杂模型之间的匹配,以及有用和无用任务特征的区分,这些方法存在过拟合问题。为了解决这一问题,我们提出了一种基于梯度相似度(AMGS)的自适应元学习器,以提高模型对新任务的泛化能力。具体而言,本文提出的AMGS从两个方面缓解了过拟合问题:(1)在内环中通过自监督辅助任务获取样本的潜在语义表示并提高模型泛化;(2)在外环中利用自适应元学习器通过梯度相似性对基础学习器获得的梯度添加约束。此外,我们还系统地分析了正则化对整个框架的影响。几个基准测试的实验结果表明,与最先进的基于优化的元学习方法相比,所提出的AMGS方法持续提高了少量文本分类性能。代码可从https://github.com/Tianyi-Lei获得。
{"title":"Adaptive Meta-learner via Gradient Similarity for Few-shot Text Classification","authors":"Tianyi Lei, Honghui Hu, Qiaoyang Luo, Dezhong Peng, Xu Wang","doi":"10.48550/arXiv.2209.04702","DOIUrl":"https://doi.org/10.48550/arXiv.2209.04702","url":null,"abstract":"Few-shot text classification aims to classify the text under the few-shot scenario. Most of the previous methods adopt optimization-based meta learning to obtain task distribution. However, due to the neglect of matching between the few amount of samples and complicated models, as well as the distinction between useful and useless task features, these methods suffer from the overfitting issue. To address this issue, we propose a novel Adaptive Meta-learner via Gradient Similarity (AMGS) method to improve the model generalization ability to a new task. Specifically, the proposed AMGS alleviates the overfitting based on two aspects: (i) acquiring the potential semantic representation of samples and improving model generalization through the self-supervised auxiliary task in the inner loop, (ii) leveraging the adaptive meta-learner via gradient similarity to add constraints on the gradient obtained by base-learner in the outer loop. Moreover, we make a systematic analysis of the influence of regularization on the entire framework. Experimental results on several benchmarks demonstrate that the proposed AMGS consistently improves few-shot text classification performance compared with the state-of-the-art optimization-based meta-learning approaches. The code is available at: https://github.com/Tianyi-Lei.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"56 1","pages":"4873-4882"},"PeriodicalIF":0.0,"publicationDate":"2022-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74457308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
A Survey in Automatic Irony Processing: Linguistic, Cognitive, and Multi-X Perspectives 反语自动加工研究进展:语言、认知和多视角
Pub Date : 2022-09-10 DOI: 10.48550/arXiv.2209.04712
Qingcheng Zeng, Anran Li
Irony is a ubiquitous figurative language in daily communication. Previously, many researchers have approached irony from linguistic, cognitive science, and computational aspects. Recently, some progress have been witnessed in automatic irony processing due to the rapid development in deep neural models in natural language processing (NLP). In this paper, we will provide a comprehensive overview of computational irony, insights from linguisic theory and cognitive science, as well as its interactions with downstream NLP tasks and newly proposed multi-X irony processing perspectives.
反讽是日常交际中普遍存在的一种比喻性语言。以前,许多研究人员从语言学、认知科学和计算方面研究反讽。近年来,由于深度神经模型在自然语言处理(NLP)中的快速发展,在反语自动处理方面取得了一些进展。在本文中,我们将全面概述计算反语,从语言学理论和认知科学的见解,以及它与下游NLP任务的相互作用和新提出的多重反语处理观点。
{"title":"A Survey in Automatic Irony Processing: Linguistic, Cognitive, and Multi-X Perspectives","authors":"Qingcheng Zeng, Anran Li","doi":"10.48550/arXiv.2209.04712","DOIUrl":"https://doi.org/10.48550/arXiv.2209.04712","url":null,"abstract":"Irony is a ubiquitous figurative language in daily communication. Previously, many researchers have approached irony from linguistic, cognitive science, and computational aspects. Recently, some progress have been witnessed in automatic irony processing due to the rapid development in deep neural models in natural language processing (NLP). In this paper, we will provide a comprehensive overview of computational irony, insights from linguisic theory and cognitive science, as well as its interactions with downstream NLP tasks and newly proposed multi-X irony processing perspectives.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"249 1","pages":"824-836"},"PeriodicalIF":0.0,"publicationDate":"2022-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82907223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Multi-Document Scientific Summarization from a Knowledge Graph-Centric View 以知识图为中心的多文献科学摘要
Pub Date : 2022-09-09 DOI: 10.48550/arXiv.2209.04319
Pancheng Wang, Shasha Li, Kunyuan Pang, Liangliang He, Dong Li, Jintao Tang, Ting Wang
Multi-Document Scientific Summarization (MDSS) aims to produce coherent and concise summaries for clusters of topic-relevant scientific papers. This task requires precise understanding of paper content and accurate modeling of cross-paper relationships. Knowledge graphs convey compact and interpretable structured information for documents, which makes them ideal for content modeling and relationship modeling. In this paper, we present KGSum, an MDSS model centred on knowledge graphs during both the encoding and decoding process. Specifically, in the encoding process, two graph-based modules are proposed to incorporate knowledge graph information into paper encoding, while in the decoding process, we propose a two-stage decoder by first generating knowledge graph information of summary in the form of descriptive sentences, followed by generating the final summary. Empirical results show that the proposed architecture brings substantial improvements over baselines on the Multi-Xscience dataset.
多文件科学摘要(MDSS)旨在为与主题相关的科学论文集群提供连贯和简明的摘要。这项任务要求对论文内容有精确的理解,并对跨论文关系进行准确的建模。知识图为文档传递紧凑且可解释的结构化信息,这使得它们非常适合于内容建模和关系建模。在本文中,我们提出了KGSum,一个在编码和解码过程中都以知识图为中心的MDSS模型。具体来说,在编码过程中,我们提出了两个基于图的模块,将知识图信息整合到纸面编码中,而在解码过程中,我们提出了一个两阶段解码器,首先以描述性句子的形式生成摘要的知识图信息,然后生成最终的摘要。实验结果表明,所提出的架构在Multi-Xscience数据集上带来了实质性的改进。
{"title":"Multi-Document Scientific Summarization from a Knowledge Graph-Centric View","authors":"Pancheng Wang, Shasha Li, Kunyuan Pang, Liangliang He, Dong Li, Jintao Tang, Ting Wang","doi":"10.48550/arXiv.2209.04319","DOIUrl":"https://doi.org/10.48550/arXiv.2209.04319","url":null,"abstract":"Multi-Document Scientific Summarization (MDSS) aims to produce coherent and concise summaries for clusters of topic-relevant scientific papers. This task requires precise understanding of paper content and accurate modeling of cross-paper relationships. Knowledge graphs convey compact and interpretable structured information for documents, which makes them ideal for content modeling and relationship modeling. In this paper, we present KGSum, an MDSS model centred on knowledge graphs during both the encoding and decoding process. Specifically, in the encoding process, two graph-based modules are proposed to incorporate knowledge graph information into paper encoding, while in the decoding process, we propose a two-stage decoder by first generating knowledge graph information of summary in the form of descriptive sentences, followed by generating the final summary. Empirical results show that the proposed architecture brings substantial improvements over baselines on the Multi-Xscience dataset.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"138 1","pages":"6222-6233"},"PeriodicalIF":0.0,"publicationDate":"2022-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79842419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Joint Alignment of Multi-Task Feature and Label Spaces for Emotion Cause Pair Extraction 面向情感原因对提取的多任务特征和标签空间联合对齐
Pub Date : 2022-09-09 DOI: 10.48550/arXiv.2209.04112
Shunjie Chen, Xiaochuan Shi, Jingye Li, Shengqiong Wu, Hao Fei, Fei Li, Donghong Ji
Emotion cause pair extraction (ECPE), as one of the derived subtasks of emotion cause analysis (ECA), shares rich inter-related features with emotion extraction (EE) and cause extraction (CE). Therefore EE and CE are frequently utilized as auxiliary tasks for better feature learning, modeled via multi-task learning (MTL) framework by prior works to achieve state-of-the-art (SoTA) ECPE results. However, existing MTL-based methods either fail to simultaneously model the specific features and the interactive feature in between, or suffer from the inconsistency of label prediction. In this work, we consider addressing the above challenges for improving ECPE by performing two alignment mechanisms with a novel Aˆ2Net model. We first propose a feature-task alignment to explicitly model the specific emotion-&cause-specific features and the shared interactive feature. Besides, an inter-task alignment is implemented, in which the label distance between the ECPE and the combinations of EE&CE are learned to be narrowed for better label consistency. Evaluations of benchmarks show that our methods outperform current best-performing systems on all ECA subtasks. Further analysis proves the importance of our proposed alignment mechanisms for the task.
情感原因对提取(ECPE)作为情感原因分析(ECA)的衍生子任务之一,与情感提取(EE)和原因提取(CE)具有丰富的相互关联特征。因此,EE和CE经常被用作更好的特征学习的辅助任务,通过多任务学习(MTL)框架进行建模,以获得最先进的(SoTA) ECPE结果。然而,现有的基于mtl的方法要么无法同时对特定特征和两者之间的交互特征进行建模,要么存在标签预测不一致的问题。在这项工作中,我们考虑通过使用一种新的awh2net模型执行两种对齐机制来解决上述改进ECPE的挑战。我们首先提出了一个特征-任务对齐来明确地建模特定的情感和原因特定的特征和共享的交互特征。此外,该算法还实现了任务间对齐,即学习缩小ECPE与EE&CE组合之间的标签距离,以获得更好的标签一致性。对基准的评估表明,我们的方法在所有ECA子任务上都优于当前表现最好的系统。进一步的分析证明了我们提出的校准机制对任务的重要性。
{"title":"Joint Alignment of Multi-Task Feature and Label Spaces for Emotion Cause Pair Extraction","authors":"Shunjie Chen, Xiaochuan Shi, Jingye Li, Shengqiong Wu, Hao Fei, Fei Li, Donghong Ji","doi":"10.48550/arXiv.2209.04112","DOIUrl":"https://doi.org/10.48550/arXiv.2209.04112","url":null,"abstract":"Emotion cause pair extraction (ECPE), as one of the derived subtasks of emotion cause analysis (ECA), shares rich inter-related features with emotion extraction (EE) and cause extraction (CE). Therefore EE and CE are frequently utilized as auxiliary tasks for better feature learning, modeled via multi-task learning (MTL) framework by prior works to achieve state-of-the-art (SoTA) ECPE results. However, existing MTL-based methods either fail to simultaneously model the specific features and the interactive feature in between, or suffer from the inconsistency of label prediction. In this work, we consider addressing the above challenges for improving ECPE by performing two alignment mechanisms with a novel Aˆ2Net model. We first propose a feature-task alignment to explicitly model the specific emotion-&cause-specific features and the shared interactive feature. Besides, an inter-task alignment is implemented, in which the label distance between the ECPE and the combinations of EE&CE are learned to be narrowed for better label consistency. Evaluations of benchmarks show that our methods outperform current best-performing systems on all ECA subtasks. Further analysis proves the importance of our proposed alignment mechanisms for the task.","PeriodicalId":91381,"journal":{"name":"Proceedings of COLING. International Conference on Computational Linguistics","volume":"25 1","pages":"6955-6965"},"PeriodicalIF":0.0,"publicationDate":"2022-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80798533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
期刊
Proceedings of COLING. International Conference on Computational Linguistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1