首页 > 最新文献

Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing最新文献

英文 中文
Pair-Based Joint Encoding with Relational Graph Convolutional Networks for Emotion-Cause Pair Extraction 基于关系图卷积网络的情感原因对联合编码
Junlong Liu, Xichen Shang, Qianli Ma
Emotion-cause pair extraction (ECPE) aims to extract emotion clauses and corresponding cause clauses, which have recently received growing attention. Previous methods sequentially encode features with a specified order. They first encode the emotion and cause features for clause extraction and then combine them for pair extraction. This lead to an imbalance in inter-task feature interaction where features extracted later have no direct contact with the former. To address this issue, we propose a novel **P**air-**B**ased **J**oint **E**ncoding (**PBJE**) network, which generates pairs and clauses features simultaneously in a joint feature encoding manner to model the causal relationship in clauses. PBJE can balance the information flow among emotion clauses, cause clauses and pairs. From a multi-relational perspective, we construct a heterogeneous undirected graph and apply the Relational Graph Convolutional Network (RGCN) to capture the multiplex relationship between clauses and the relationship between pairs and clauses. Experimental results show that PBJE achieves state-of-the-art performance on the Chinese benchmark corpus.
情感-原因对提取(ECPE)旨在提取情感子句和相应的原因子句,近年来受到越来越多的关注。以前的方法按照指定的顺序对特征进行顺序编码。他们首先对情感和原因特征进行编码进行子句提取,然后将它们组合起来进行对提取。这导致了任务间特征交互的不平衡,即后提取的特征与前提取的特征没有直接联系。为了解决这一问题,我们提出了一种新的基于**P**air-**B* based **J** point **E**ncoding (**PBJE**)网络,该网络以联合特征编码的方式同时生成对和子句特征,对子句中的因果关系进行建模。PBJE能够平衡情感子句、原因子句和对之间的信息流。从多关系的角度出发,构造了一个异构无向图,并应用关系图卷积网络(RGCN)捕捉子句之间以及对与子句之间的多重关系。实验结果表明,PBJE在中文基准语料库上达到了最先进的性能。
{"title":"Pair-Based Joint Encoding with Relational Graph Convolutional Networks for Emotion-Cause Pair Extraction","authors":"Junlong Liu, Xichen Shang, Qianli Ma","doi":"10.48550/arXiv.2212.01844","DOIUrl":"https://doi.org/10.48550/arXiv.2212.01844","url":null,"abstract":"Emotion-cause pair extraction (ECPE) aims to extract emotion clauses and corresponding cause clauses, which have recently received growing attention. Previous methods sequentially encode features with a specified order. They first encode the emotion and cause features for clause extraction and then combine them for pair extraction. This lead to an imbalance in inter-task feature interaction where features extracted later have no direct contact with the former. To address this issue, we propose a novel **P**air-**B**ased **J**oint **E**ncoding (**PBJE**) network, which generates pairs and clauses features simultaneously in a joint feature encoding manner to model the causal relationship in clauses. PBJE can balance the information flow among emotion clauses, cause clauses and pairs. From a multi-relational perspective, we construct a heterogeneous undirected graph and apply the Relational Graph Convolutional Network (RGCN) to capture the multiplex relationship between clauses and the relationship between pairs and clauses. Experimental results show that PBJE achieves state-of-the-art performance on the Chinese benchmark corpus.","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"27 1","pages":"5339-5351"},"PeriodicalIF":0.0,"publicationDate":"2022-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82311837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Constructing Highly Inductive Contexts for Dialogue Safety through Controllable Reverse Generation 通过可控逆向生成构建高度归纳的对话安全语境
Zhexin Zhang, Jiale Cheng, Hao Sun, Jiawen Deng, Fei Mi, Yasheng Wang, Lifeng Shang, Minlie Huang
Large pretrained language models can easily produce toxic or biased content, which is prohibitive for practical use. In order to detect such toxic generations, existing methods rely on templates, real-world data extraction, crowdsourcing workers, or automatic generation to construct adversarial contexts that are likely to induce toxic generations. However, what type of context is more likely to induce unsafe responses is still under-explored. In this paper, we identify that context toxicity and context category (e.g., textit{profanity}, textit{insult}, textit{drugs}, etc.) are two important factors to cause safety issues in response generation. Hence, we propose a method called emph{reverse generation} to construct adversarial contexts conditioned on a given response, with the flexibility to control category, toxicity level, and inductivity of the generated contexts. Via reverse generation, we augment the existing BAD dataset and construct a new dataset BAD+ which contains more than 120K diverse and highly inductive contexts in 12 categories. We test three popular pretrained dialogue models (Blender, DialoGPT, and Plato2) and find that BAD+ can largely expose their safety problems. Furthermore, we show that BAD+ can greatly enhance the safety of generation and reveal the key factors of safety improvement. Our code and dataset is available at url{https://github.com/thu-coai/Reverse_Generation}.
大型预训练语言模型很容易产生有害或有偏见的内容,这不利于实际使用。为了检测这些有毒代,现有的方法依赖于模板、现实世界的数据提取、众包工人或自动生成来构建可能诱导有毒代的对抗性环境。然而,哪种类型的环境更有可能引起不安全的反应仍未得到充分探讨。在本文中,我们确定上下文毒性和上下文类别(例如,textit{亵渎},textit{侮辱},textit{药物}等)是导致响应生成安全问题的两个重要因素。因此,我们提出了一种称为emph{反向生成}的方法来构建基于给定响应的对抗性上下文,并具有控制生成上下文的类别、毒性水平和归纳性的灵活性。通过反向生成,我们增强了现有的BAD数据集,并构建了一个新的数据集BAD+,该数据集包含12个类别中超过120K个不同且高度归纳的上下文。我们测试了三种流行的预训练对话模型(Blender、DialoGPT和Plato2),发现BAD+可以在很大程度上暴露它们的安全问题。此外,我们还发现BAD+可以大大提高发电安全性,并揭示了安全改进的关键因素。我们的代码和数据集可在url{https://github.com/thu-coai/Reverse_Generation}上获得。
{"title":"Constructing Highly Inductive Contexts for Dialogue Safety through Controllable Reverse Generation","authors":"Zhexin Zhang, Jiale Cheng, Hao Sun, Jiawen Deng, Fei Mi, Yasheng Wang, Lifeng Shang, Minlie Huang","doi":"10.48550/arXiv.2212.01810","DOIUrl":"https://doi.org/10.48550/arXiv.2212.01810","url":null,"abstract":"Large pretrained language models can easily produce toxic or biased content, which is prohibitive for practical use. In order to detect such toxic generations, existing methods rely on templates, real-world data extraction, crowdsourcing workers, or automatic generation to construct adversarial contexts that are likely to induce toxic generations. However, what type of context is more likely to induce unsafe responses is still under-explored. In this paper, we identify that context toxicity and context category (e.g., textit{profanity}, textit{insult}, textit{drugs}, etc.) are two important factors to cause safety issues in response generation. Hence, we propose a method called emph{reverse generation} to construct adversarial contexts conditioned on a given response, with the flexibility to control category, toxicity level, and inductivity of the generated contexts. Via reverse generation, we augment the existing BAD dataset and construct a new dataset BAD+ which contains more than 120K diverse and highly inductive contexts in 12 categories. We test three popular pretrained dialogue models (Blender, DialoGPT, and Plato2) and find that BAD+ can largely expose their safety problems. Furthermore, we show that BAD+ can greatly enhance the safety of generation and reveal the key factors of safety improvement. Our code and dataset is available at url{https://github.com/thu-coai/Reverse_Generation}.","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"49 1","pages":"3684-3697"},"PeriodicalIF":0.0,"publicationDate":"2022-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87400812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
T-STAR: Truthful Style Transfer using AMR Graph as Intermediate Representation T-STAR:用AMR图作为中间表示的真实风格迁移
Anubhav Jangra, Preksha Nema, A. Raghuveer
Unavailability of parallel corpora for training text style transfer (TST) models is a very challenging yet common scenario. Also, TST models implicitly need to preserve the content while transforming a source sentence into the target style. To tackle these problems, an intermediate representation is often constructed that is devoid of style while still preserving the meaning of the source sentence. In this work, we study the usefulness of Abstract Meaning Representation (AMR) graph as the intermediate style agnostic representation. We posit that semantic notations like AMR are a natural choice for an intermediate representation. Hence, we propose T-STAR: a model comprising of two components, text-to-AMR encoder and a AMR-to-text decoder. We propose several modeling improvements to enhance the style agnosticity of the generated AMR. To the best of our knowledge, T-STAR is the first work that uses AMR as an intermediate representation for TST. With thorough experimental evaluation we show T-STAR significantly outperforms state of the art techniques by achieving on an average 15.2% higher content preservation with negligible loss (~3%) in style accuracy. Through detailed human evaluation with 90,000 ratings, we also show that T-STAR has upto 50% lesser hallucinations compared to state of the art TST models.
无法获得用于训练文本风格迁移(TST)模型的并行语料库是一个非常具有挑战性但又很常见的情况。此外,TST模型隐式地需要在将源句子转换为目标风格时保留内容。为了解决这些问题,通常构建一个没有风格的中间表示,同时仍然保留源句子的含义。在这项工作中,我们研究了抽象意义表示(AMR)图作为中间风格不可知论表示的有效性。我们假设像AMR这样的语义符号是中间表示的自然选择。因此,我们提出了T-STAR:一个由两个组件组成的模型,文本到amr编码器和amr到文本解码器。我们提出了几个建模改进,以增强生成的AMR的风格不可知性。据我们所知,T-STAR是第一个使用AMR作为TST的中间表示的工作。经过彻底的实验评估,我们表明T-STAR显著优于最先进的技术,平均提高了15.2%的内容保存,而风格准确性的损失可以忽略不计(约3%)。通过对90,000个评分的详细人类评估,我们还表明,与最先进的TST模型相比,T-STAR的幻觉减少了50%。
{"title":"T-STAR: Truthful Style Transfer using AMR Graph as Intermediate Representation","authors":"Anubhav Jangra, Preksha Nema, A. Raghuveer","doi":"10.48550/arXiv.2212.01667","DOIUrl":"https://doi.org/10.48550/arXiv.2212.01667","url":null,"abstract":"Unavailability of parallel corpora for training text style transfer (TST) models is a very challenging yet common scenario. Also, TST models implicitly need to preserve the content while transforming a source sentence into the target style. To tackle these problems, an intermediate representation is often constructed that is devoid of style while still preserving the meaning of the source sentence. In this work, we study the usefulness of Abstract Meaning Representation (AMR) graph as the intermediate style agnostic representation. We posit that semantic notations like AMR are a natural choice for an intermediate representation. Hence, we propose T-STAR: a model comprising of two components, text-to-AMR encoder and a AMR-to-text decoder. We propose several modeling improvements to enhance the style agnosticity of the generated AMR. To the best of our knowledge, T-STAR is the first work that uses AMR as an intermediate representation for TST. With thorough experimental evaluation we show T-STAR significantly outperforms state of the art techniques by achieving on an average 15.2% higher content preservation with negligible loss (~3%) in style accuracy. Through detailed human evaluation with 90,000 ratings, we also show that T-STAR has upto 50% lesser hallucinations compared to state of the art TST models.","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"27 1","pages":"8805-8825"},"PeriodicalIF":0.0,"publicationDate":"2022-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84196834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Modeling Label Correlations for Ultra-Fine Entity Typing with Neural Pairwise Conditional Random Field 基于神经对条件随机场的超精细实体分类标签关联建模
Chengyue Jiang, Yong Jiang, Weiqi Wu, Pengjun Xie, Kewei Tu
Ultra-fine entity typing (UFET) aims to predict a wide range of type phrases that correctly describe the categories of a given entity mention in a sentence. Most recent works infer each entity type independently, ignoring the correlations between types, e.g., when an entity is inferred as a president, it should also be a politician and a leader. To this end, we use an undirected graphical model called pairwise conditional random field (PCRF) to formulate the UFET problem, in which the type variables are not only unarily influenced by the input but also pairwisely relate to all the other type variables. We use various modern backbones for entity typing to compute unary potentials, and derive pairwise potentials from type phrase representations that both capture prior semantic information and facilitate accelerated inference. We use mean-field variational inference for efficient type inference on very large type sets and unfold it as a neural network module to enable end-to-end training. Experiments on UFET show that the Neural-PCRF consistently outperforms its backbones with little cost and results in a competitive performance against cross-encoder based SOTA while being thousands of times faster. We also find Neural-PCRF effective on a widely used fine-grained entity typing dataset with a smaller type set. We pack Neural-PCRF as a network module that can be plugged onto multi-label type classifiers with ease and release it in .
超精细实体类型(uet)旨在预测广泛的类型短语,正确描述句子中提到的给定实体的类别。最近的大多数作品都独立地推断出每个实体类型,忽略了类型之间的相关性,例如,当一个实体被推断为总统时,它也应该是一个政治家和领导者。为此,我们使用一种称为成对条件随机场(PCRF)的无向图形模型来表述uet问题,其中类型变量不仅受到输入的单一影响,而且还与所有其他类型变量成对相关。我们使用各种现代实体类型主干来计算一元势,并从类型短语表示中获得两两势,这既捕获了先验语义信息,又促进了加速推理。我们使用平均场变分推理对非常大的类型集进行有效的类型推理,并将其展开为一个神经网络模块,以实现端到端的训练。在uet上的实验表明,Neural-PCRF以很少的成本持续优于其骨干,并且在与基于交叉编码器的SOTA的竞争性能中具有竞争力,同时速度快数千倍。我们还发现Neural-PCRF在广泛使用的细粒度实体类型数据集上具有较小的类型集。我们将Neural-PCRF打包为一个网络模块,可以轻松地插入到多标签类型分类器中并释放它。
{"title":"Modeling Label Correlations for Ultra-Fine Entity Typing with Neural Pairwise Conditional Random Field","authors":"Chengyue Jiang, Yong Jiang, Weiqi Wu, Pengjun Xie, Kewei Tu","doi":"10.48550/arXiv.2212.01581","DOIUrl":"https://doi.org/10.48550/arXiv.2212.01581","url":null,"abstract":"Ultra-fine entity typing (UFET) aims to predict a wide range of type phrases that correctly describe the categories of a given entity mention in a sentence. Most recent works infer each entity type independently, ignoring the correlations between types, e.g., when an entity is inferred as a president, it should also be a politician and a leader. To this end, we use an undirected graphical model called pairwise conditional random field (PCRF) to formulate the UFET problem, in which the type variables are not only unarily influenced by the input but also pairwisely relate to all the other type variables. We use various modern backbones for entity typing to compute unary potentials, and derive pairwise potentials from type phrase representations that both capture prior semantic information and facilitate accelerated inference. We use mean-field variational inference for efficient type inference on very large type sets and unfold it as a neural network module to enable end-to-end training. Experiments on UFET show that the Neural-PCRF consistently outperforms its backbones with little cost and results in a competitive performance against cross-encoder based SOTA while being thousands of times faster. We also find Neural-PCRF effective on a widely used fine-grained entity typing dataset with a smaller type set. We pack Neural-PCRF as a network module that can be plugged onto multi-label type classifiers with ease and release it in .","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"303 1","pages":"6836-6847"},"PeriodicalIF":0.0,"publicationDate":"2022-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77494631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Subword-Delimited Downsampling for Better Character-Level Translation 子词分隔的下采样,以获得更好的字符级翻译
Lukas Edman, Antonio Toral, Gertjan van Noord
Subword-level models have been the dominant paradigm in NLP. However, character-level models have the benefit of seeing each character individually, providing the model with more detailed information that ultimately could lead to better models. Recent works have shown character-level models to be competitive with subword models, but costly in terms of time and computation. Character-level models with a downsampling component alleviate this, but at the cost of quality, particularly for machine translation. This work analyzes the problems of previous downsampling methods and introduces a novel downsampling method which is informed by subwords. This new downsampling method not only outperforms existing downsampling methods, showing that downsampling characters can be done without sacrificing quality, but also leads to promising performance compared to subword models for translation.
子词级模型一直是自然语言处理的主流范式。然而,角色级模型的好处是可以单独看到每个角色,为模型提供更详细的信息,最终可以生成更好的模型。最近的研究表明,字符级模型可以与子词模型竞争,但在时间和计算方面代价高昂。带有下采样组件的字符级模型缓解了这一点,但代价是质量,特别是对于机器翻译。本文分析了以往下采样方法存在的问题,提出了一种新的基于子词的下采样方法。这种新的下采样方法不仅优于现有的下采样方法,表明下采样字符可以在不牺牲质量的情况下完成,而且与翻译的子词模型相比,它的性能也很好。
{"title":"Subword-Delimited Downsampling for Better Character-Level Translation","authors":"Lukas Edman, Antonio Toral, Gertjan van Noord","doi":"10.48550/arXiv.2212.01304","DOIUrl":"https://doi.org/10.48550/arXiv.2212.01304","url":null,"abstract":"Subword-level models have been the dominant paradigm in NLP. However, character-level models have the benefit of seeing each character individually, providing the model with more detailed information that ultimately could lead to better models. Recent works have shown character-level models to be competitive with subword models, but costly in terms of time and computation. Character-level models with a downsampling component alleviate this, but at the cost of quality, particularly for machine translation. This work analyzes the problems of previous downsampling methods and introduces a novel downsampling method which is informed by subwords. This new downsampling method not only outperforms existing downsampling methods, showing that downsampling characters can be done without sacrificing quality, but also leads to promising performance compared to subword models for translation.","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"58 1","pages":"981-992"},"PeriodicalIF":0.0,"publicationDate":"2022-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81634045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Semantic Role Labeling Meets Definition Modeling: Using Natural Language to Describe Predicate-Argument Structures 语义角色标注符合定义建模:使用自然语言描述谓词-参数结构
Simone Conia, Edoardo Barba, Alessandro Sciré, Roberto Navigli
One of the common traits of past and present approaches for Semantic Role Labeling (SRL) is that they rely upon discrete labels drawn from a predefined linguistic inventory to classify predicate senses and their arguments. However, we argue this need not be the case. In this paper, we present an approach that leverages Definition Modeling to introduce a generalized formulation of SRL as the task of describing predicate-argument structures using natural language definitions instead of discrete labels. Our novel formulation takes a first step towards placing interpretability and flexibility foremost, and yet our experiments and analyses on PropBank-style and FrameNet-style, dependency-based and span-based SRL also demonstrate that a flexible model with an interpretable output does not necessarily come at the expense of performance. We release our software for research purposes at https://github.com/SapienzaNLP/dsrl.
过去和现在的语义角色标注(SRL)方法的一个共同特点是,它们依赖于从预定义的语言清单中提取的离散标签来对谓词意义及其参数进行分类。然而,我们认为事实并非如此。在本文中,我们提出了一种利用定义建模来引入SRL的广义公式的方法,作为使用自然语言定义而不是离散标签描述谓词-参数结构的任务。我们的新公式向将可解释性和灵活性放在首位迈出了第一步,然而我们对propbank风格和framework风格、基于依赖和基于跨度的SRL的实验和分析也表明,具有可解释性输出的灵活模型并不一定以牺牲性能为代价。我们发布我们的软件用于研究目的在https://github.com/SapienzaNLP/dsrl。
{"title":"Semantic Role Labeling Meets Definition Modeling: Using Natural Language to Describe Predicate-Argument Structures","authors":"Simone Conia, Edoardo Barba, Alessandro Sciré, Roberto Navigli","doi":"10.48550/arXiv.2212.01094","DOIUrl":"https://doi.org/10.48550/arXiv.2212.01094","url":null,"abstract":"One of the common traits of past and present approaches for Semantic Role Labeling (SRL) is that they rely upon discrete labels drawn from a predefined linguistic inventory to classify predicate senses and their arguments. However, we argue this need not be the case. In this paper, we present an approach that leverages Definition Modeling to introduce a generalized formulation of SRL as the task of describing predicate-argument structures using natural language definitions instead of discrete labels. Our novel formulation takes a first step towards placing interpretability and flexibility foremost, and yet our experiments and analyses on PropBank-style and FrameNet-style, dependency-based and span-based SRL also demonstrate that a flexible model with an interpretable output does not necessarily come at the expense of performance. We release our software for research purposes at https://github.com/SapienzaNLP/dsrl.","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"32 1","pages":"4253-4270"},"PeriodicalIF":0.0,"publicationDate":"2022-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83818069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
NarraSum: A Large-Scale Dataset for Abstractive Narrative Summarization narasum:用于抽象叙述摘要的大规模数据集
Chao Zhao, Faeze Brahman, Kaiqiang Song, Wenlin Yao, Dian Yu, Snigdha Chaturvedi
Narrative summarization aims to produce a distilled version of a narrative to describe its most salient events and characters. Summarizing a narrative is challenging as it requires an understanding of event causality and character behaviors. To encourage research in this direction, we propose NarraSum, a large-scale narrative summarization dataset. It contains 122K narrative documents, which are collected from plot descriptions of movies and TV episodes with diverse genres, and their corresponding abstractive summaries. Experiments show that there is a large performance gap between humans and the state-of-the-art summarization models on NarraSum. We hope that this dataset will promote future research in summarization, as well as broader studies of natural language understanding and generation. The dataset is available at https://github.com/zhaochaocs/narrasum.
叙述摘要的目的是提炼出一篇叙述的精华,以描述其中最突出的事件和人物。总结一个故事是很有挑战性的,因为它需要理解事件的因果关系和角色的行为。为了鼓励这方面的研究,我们提出了一个大型叙事摘要数据集narasum。它包含122K个叙事文件,这些文件收集了不同类型的电影和电视剧集的情节描述及其相应的抽象摘要。实验表明,在narasum上,人类和最先进的总结模型之间存在很大的性能差距。我们希望这个数据集能够促进未来的总结研究,以及更广泛的自然语言理解和生成研究。该数据集可在https://github.com/zhaochaocs/narrasum上获得。
{"title":"NarraSum: A Large-Scale Dataset for Abstractive Narrative Summarization","authors":"Chao Zhao, Faeze Brahman, Kaiqiang Song, Wenlin Yao, Dian Yu, Snigdha Chaturvedi","doi":"10.48550/arXiv.2212.01476","DOIUrl":"https://doi.org/10.48550/arXiv.2212.01476","url":null,"abstract":"Narrative summarization aims to produce a distilled version of a narrative to describe its most salient events and characters. Summarizing a narrative is challenging as it requires an understanding of event causality and character behaviors. To encourage research in this direction, we propose NarraSum, a large-scale narrative summarization dataset. It contains 122K narrative documents, which are collected from plot descriptions of movies and TV episodes with diverse genres, and their corresponding abstractive summaries. Experiments show that there is a large performance gap between humans and the state-of-the-art summarization models on NarraSum. We hope that this dataset will promote future research in summarization, as well as broader studies of natural language understanding and generation. The dataset is available at https://github.com/zhaochaocs/narrasum.","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"66 1","pages":"182-197"},"PeriodicalIF":0.0,"publicationDate":"2022-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76995352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Dynamic Augmentation Data Selection for Few-shot Text Classification. 为少量文本分类选择动态增强数据
Guangliang Liu, Owen Yuan, Lifeng Jin, Jiayu Zhou

Data augmentation has been a popular method for fine-tuning pre-trained language models to increase model robustness and performance. With augmentation data coming from modifying gold train data (in-sample augmentation) or being harvested from general domain unlabeled data (out-of-sample augmentation), the quality of such data is the key to successful fine-tuning. In this paper, we propose a dynamic data selection method to select effective augmentation data from different augmentation sources according to the model's learning stage, by identifying a set of augmentation samples that optimally facilitates the learning process of the most current model. The method firstly filters out augmentation samples with noisy pseudo labels through a curriculum learning strategy, then estimates the effectiveness of reserved augmentation data by its influence scores on the current model at every update, allowing the data selection process tightly tailored to model parameters. And the two-stage augmentation strategy considers in-sample augmentation and out-of-sample augmentation in different learning stages. Experiments with both kinds of augmentation data on a variety of sentence classification tasks show that our method outperforms strong baselines, proving the effectiveness of our method. Analysis confirms the dynamic nature of the data effectiveness and the importance of model learning stages in utilization of augmentation data.

数据扩增一直是微调预训练语言模型以提高模型稳健性和性能的常用方法。扩增数据来自修改黄金训练数据(样本内扩增)或从一般领域的无标记数据中获取(样本外扩增),这些数据的质量是微调成功的关键。在本文中,我们提出了一种动态数据选择方法,根据模型的学习阶段,从不同的增强来源中选择有效的增强数据,通过识别一组增强样本来优化当前模型的学习过程。该方法首先通过课程学习策略过滤掉带有噪声伪标签的增强样本,然后在每次更新时通过其对当前模型的影响分数来估计保留的增强数据的有效性,从而使数据选择过程与模型参数紧密契合。两阶段增强策略在不同的学习阶段考虑了样本内增强和样本外增强。使用这两种增强数据对各种句子分类任务进行的实验表明,我们的方法优于强基准,证明了我们方法的有效性。分析证实了数据有效性的动态性质以及模型学习阶段在利用增强数据方面的重要性。
{"title":"Dynamic Augmentation Data Selection for Few-shot Text Classification.","authors":"Guangliang Liu, Owen Yuan, Lifeng Jin, Jiayu Zhou","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Data augmentation has been a popular method for fine-tuning pre-trained language models to increase model robustness and performance. With augmentation data coming from modifying gold train data (in-sample augmentation) or being harvested from general domain unlabeled data (out-of-sample augmentation), the quality of such data is the key to successful fine-tuning. In this paper, we propose a dynamic data selection method to select effective augmentation data from different augmentation sources according to the model's learning stage, by identifying a set of augmentation samples that optimally facilitates the learning process of the most current model. The method firstly filters out augmentation samples with noisy pseudo labels through a curriculum learning strategy, then estimates the effectiveness of reserved augmentation data by its influence scores on the current model at every update, allowing the data selection process tightly tailored to model parameters. And the two-stage augmentation strategy considers in-sample augmentation and out-of-sample augmentation in different learning stages. Experiments with both kinds of augmentation data on a variety of sentence classification tasks show that our method outperforms strong baselines, proving the effectiveness of our method. Analysis confirms the dynamic nature of the data effectiveness and the importance of model learning stages in utilization of augmentation data.</p>","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"2022 ","pages":"4870-4881"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10097500/pdf/nihms-1888098.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9302914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analogical Math Word Problems Solving with Enhanced Problem-Solution Association 用增强的问题-解决方案关联解决类比数学单词问题
Zhenwen Liang, Jipeng Zhang, Xiangliang Zhang
Math word problem (MWP) solving is an important task in question answering which requires human-like reasoning ability. Analogical reasoning has long been used in mathematical education, as it enables students to apply common relational structures of mathematical situations to solve new problems. In this paper, we propose to build a novel MWP solver by leveraging analogical MWPs, which advance the solver’s generalization ability across different kinds of MWPs. The key idea, named analogy identification, is to associate the analogical MWP pairs in a latent space, i.e., encoding an MWP close to another analogical MWP, while leaving away from the non-analogical ones. Moreover, a solution discriminator is integrated into the MWP solver to enhance the association between an MWP and its true solution. The evaluation results verify that our proposed analogical learning strategy promotes the performance of MWP-BERT on Math23k over the state-of-the-art model Generate2Rank, with 5 times fewer parameters in the encoder. We also find that our model has a stronger generalization ability in solving difficult MWPs due to the analogical learning from easy MWPs.
数学应用题求解是问答中的一项重要任务,需要具备类似人的推理能力。类比推理在数学教育中一直被使用,因为它使学生能够应用数学情境的常见关系结构来解决新问题。在本文中,我们提出利用类比MWP构建一个新的MWP求解器,这提高了求解器在不同类型MWP之间的泛化能力。其关键思想,称为类比识别,是在潜在空间中关联类比MWP对,即编码一个接近另一个类比MWP的MWP,而远离非类比MWP。此外,在MWP求解器中集成了解判别器,增强了MWP与其真解之间的关联。评估结果验证了我们提出的类比学习策略在Math23k上比最先进的模型Generate2Rank提高了MWP-BERT的性能,编码器中的参数减少了5倍。我们还发现,由于从简单的mwp中进行类比学习,我们的模型在解决困难的mwp时具有更强的泛化能力。
{"title":"Analogical Math Word Problems Solving with Enhanced Problem-Solution Association","authors":"Zhenwen Liang, Jipeng Zhang, Xiangliang Zhang","doi":"10.48550/arXiv.2212.00837","DOIUrl":"https://doi.org/10.48550/arXiv.2212.00837","url":null,"abstract":"Math word problem (MWP) solving is an important task in question answering which requires human-like reasoning ability. Analogical reasoning has long been used in mathematical education, as it enables students to apply common relational structures of mathematical situations to solve new problems. In this paper, we propose to build a novel MWP solver by leveraging analogical MWPs, which advance the solver’s generalization ability across different kinds of MWPs. The key idea, named analogy identification, is to associate the analogical MWP pairs in a latent space, i.e., encoding an MWP close to another analogical MWP, while leaving away from the non-analogical ones. Moreover, a solution discriminator is integrated into the MWP solver to enhance the association between an MWP and its true solution. The evaluation results verify that our proposed analogical learning strategy promotes the performance of MWP-BERT on Math23k over the state-of-the-art model Generate2Rank, with 5 times fewer parameters in the encoder. We also find that our model has a stronger generalization ability in solving difficult MWPs due to the analogical learning from easy MWPs.","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"44 1","pages":"9454-9464"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79437917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
IRRGN: An Implicit Relational Reasoning Graph Network for Multi-turn Response Selection 多回合响应选择的隐式关系推理图网络
Jing-Hui Deng, Hengwei Dai, Xuewei Guo, Yuanchen Ju, Wei Peng
The task of response selection in multi-turn dialogue is to find the best option from all candidates. In order to improve the reasoning ability of the model, previous studies pay more attention to using explicit algorithms to model the dependencies between utterances, which are deterministic, limited and inflexible. In addition, few studies consider differences between the options before and after reasoning. In this paper, we propose an Implicit Relational Reasoning Graph Network to address these issues, which consists of the Utterance Relational Reasoner (URR) and the Option Dual Comparator (ODC). URR aims to implicitly extract dependencies between utterances, as well as utterances and options, and make reasoning with relational graph convolutional networks. ODC focuses on perceiving the difference between the options through dual comparison, which can eliminate the interference of the noise options. Experimental results on two multi-turn dialogue reasoning benchmark datasets MuTual and MuTualplus show that our method significantly improves the baseline of four pre-trained language models and achieves state-of-the-art performance. The model surpasses human performance for the first time on the MuTual dataset.
在多回合对话中,回答选择的任务是从所有的候选者中找出最佳选择。为了提高模型的推理能力,以往的研究更注重使用显式算法对话语之间的依赖关系进行建模,而话语之间的依赖关系具有确定性、有限性和不灵活性。此外,很少有研究考虑推理前后选项之间的差异。在本文中,我们提出了一个隐式关系推理图网络来解决这些问题,它由话语关系推理器(URR)和选项双比较器(ODC)组成。URR旨在隐式提取话语之间的依赖关系,以及话语和选项之间的依赖关系,并使用关系图卷积网络进行推理。ODC侧重于通过双重比较感知选项之间的差异,可以消除噪声选项的干扰。在MuTual和MuTualplus两个多回合对话推理基准数据集上的实验结果表明,我们的方法显著提高了四种预训练语言模型的基线,达到了最先进的性能。该模型首次在MuTual数据集上超越了人类的表现。
{"title":"IRRGN: An Implicit Relational Reasoning Graph Network for Multi-turn Response Selection","authors":"Jing-Hui Deng, Hengwei Dai, Xuewei Guo, Yuanchen Ju, Wei Peng","doi":"10.48550/arXiv.2212.00482","DOIUrl":"https://doi.org/10.48550/arXiv.2212.00482","url":null,"abstract":"The task of response selection in multi-turn dialogue is to find the best option from all candidates. In order to improve the reasoning ability of the model, previous studies pay more attention to using explicit algorithms to model the dependencies between utterances, which are deterministic, limited and inflexible. In addition, few studies consider differences between the options before and after reasoning. In this paper, we propose an Implicit Relational Reasoning Graph Network to address these issues, which consists of the Utterance Relational Reasoner (URR) and the Option Dual Comparator (ODC). URR aims to implicitly extract dependencies between utterances, as well as utterances and options, and make reasoning with relational graph convolutional networks. ODC focuses on perceiving the difference between the options through dual comparison, which can eliminate the interference of the noise options. Experimental results on two multi-turn dialogue reasoning benchmark datasets MuTual and MuTualplus show that our method significantly improves the baseline of four pre-trained language models and achieves state-of-the-art performance. The model surpasses human performance for the first time on the MuTual dataset.","PeriodicalId":74540,"journal":{"name":"Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing","volume":"36 1","pages":"8529-8541"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90914055","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1