首页 > 最新文献

Findings (Sydney (N.S.W.)最新文献

英文 中文
Stabilized In-Context Learning with Pre-trained Language Models for Few Shot Dialogue State Tracking 用于少镜头对话状态跟踪的预训练语言模型的稳定上下文学习
Pub Date : 2023-02-12 DOI: 10.48550/arXiv.2302.05932
Derek Chen, Kun Qian, Zhou Yu
Prompt-based methods with large pre-trained language models (PLMs) have shown impressive unaided performance across many NLP tasks. These models improve even further with the addition of a few labeled in-context exemplars to guide output generation. However, for more complex tasks such as dialogue state tracking (DST), designing prompts that reliably convey the desired intent is nontrivial, leading to unstable results. Furthermore, building in-context exemplars for dialogue tasks is difficult because conversational contexts are long while model input lengths are relatively short.To overcome these issues we first adapt a meta-learning scheme to the dialogue domain which stabilizes the ability of the model to perform well under various prompts. We additionally design a novel training method to improve upon vanilla retrieval mechanisms to find ideal in-context examples. Finally, we introduce a saliency model to limit dialogue text length, allowing us to include more exemplars per query. In effect, we are able to achieve highly competitive results for few-shot DST on MultiWOZ.
具有大型预训练语言模型(PLM)的基于提示的方法在许多NLP任务中显示出令人印象深刻的独立性能。通过添加一些标记在上下文中的示例来指导输出生成,这些模型得到了进一步的改进。然而,对于更复杂的任务,如对话状态跟踪(DST),设计可靠地传达所需意图的提示是不寻常的,会导致不稳定的结果。此外,为对话任务构建上下文示例是困难的,因为对话上下文很长,而模型输入长度相对较短。为了克服这些问题,我们首先将元学习方案应用于对话域,这稳定了模型在各种提示下表现良好的能力。我们还设计了一种新的训练方法来改进普通的检索机制,以找到理想的上下文示例。最后,我们引入了一个显著性模型来限制对话文本的长度,允许我们在每个查询中包含更多的样例。实际上,我们能够在MultiWOZ上实现具有竞争力的少镜头DST结果。
{"title":"Stabilized In-Context Learning with Pre-trained Language Models for Few Shot Dialogue State Tracking","authors":"Derek Chen, Kun Qian, Zhou Yu","doi":"10.48550/arXiv.2302.05932","DOIUrl":"https://doi.org/10.48550/arXiv.2302.05932","url":null,"abstract":"Prompt-based methods with large pre-trained language models (PLMs) have shown impressive unaided performance across many NLP tasks. These models improve even further with the addition of a few labeled in-context exemplars to guide output generation. However, for more complex tasks such as dialogue state tracking (DST), designing prompts that reliably convey the desired intent is nontrivial, leading to unstable results. Furthermore, building in-context exemplars for dialogue tasks is difficult because conversational contexts are long while model input lengths are relatively short.To overcome these issues we first adapt a meta-learning scheme to the dialogue domain which stabilizes the ability of the model to perform well under various prompts. We additionally design a novel training method to improve upon vanilla retrieval mechanisms to find ideal in-context examples. Finally, we introduce a saliency model to limit dialogue text length, allowing us to include more exemplars per query. In effect, we are able to achieve highly competitive results for few-shot DST on MultiWOZ.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"1506-1519"},"PeriodicalIF":0.0,"publicationDate":"2023-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43769204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Discourse Structure Extraction from Pre-Trained and Fine-Tuned Language Models in Dialogues 基于预训练和微调语言模型的对话语篇结构提取
Pub Date : 2023-02-12 DOI: 10.48550/arXiv.2302.05895
Chuyuan Li, Patrick Huber, Wen Xiao, M. Amblard, Chloé Braud, G. Carenini
Discourse processing suffers from data sparsity, especially for dialogues. As a result, we explore approaches to infer latent discourse structures for dialogues, based on attention matrices from Pre-trained Language Models (PLMs). We investigate multiple auxiliary tasks for fine-tuning and show that the dialogue-tailored Sentence Ordering task performs best. To locate and exploit discourse information in PLMs, we propose an unsupervised and a semi-supervised method. Our proposals thereby achieve encouraging results on the STAC corpus, with F1 scores of 57.2 and 59.3 for the unsupervised and semi-supervised methods, respectively. When restricted to projective trees, our scores improved to 63.3 and 68.1.
语篇处理受到数据稀疏性的影响,尤其是在对话中。因此,我们探索了基于预训练语言模型的注意力矩阵来推断对话潜在话语结构的方法。我们研究了用于微调的多个辅助任务,并表明对话定制的句子排序任务表现最好。为了定位和利用PLM中的话语信息,我们提出了一种无监督和半监督的方法。因此,我们的建议在STAC语料库上取得了令人鼓舞的结果,无监督和半监督方法的F1得分分别为57.2和59.3。当仅限于投影树时,我们的得分分别提高到63.3和68.1。
{"title":"Discourse Structure Extraction from Pre-Trained and Fine-Tuned Language Models in Dialogues","authors":"Chuyuan Li, Patrick Huber, Wen Xiao, M. Amblard, Chloé Braud, G. Carenini","doi":"10.48550/arXiv.2302.05895","DOIUrl":"https://doi.org/10.48550/arXiv.2302.05895","url":null,"abstract":"Discourse processing suffers from data sparsity, especially for dialogues. As a result, we explore approaches to infer latent discourse structures for dialogues, based on attention matrices from Pre-trained Language Models (PLMs). We investigate multiple auxiliary tasks for fine-tuning and show that the dialogue-tailored Sentence Ordering task performs best. To locate and exploit discourse information in PLMs, we propose an unsupervised and a semi-supervised method. Our proposals thereby achieve encouraging results on the STAC corpus, with F1 scores of 57.2 and 59.3 for the unsupervised and semi-supervised methods, respectively. When restricted to projective trees, our scores improved to 63.3 and 68.1.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"2517-2534"},"PeriodicalIF":0.0,"publicationDate":"2023-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43033428","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Divergence-Based Domain Transferability for Zero-Shot Classification 基于散度的零射分类领域可转移性
Pub Date : 2023-02-11 DOI: 10.48550/arXiv.2302.05735
Alexander Pugantsov, R. McCreadie
Transferring learned patterns from pretrained neural language models has been shown to significantly improve effectiveness across a variety of language-based tasks, meanwhile further tuning on intermediate tasks has been demonstrated to provide additional performance benefits, provided the intermediate task is sufficiently related to the target task. However, how to identify related tasks is an open problem, and brute-force searching effective task combinations is prohibitively expensive. Hence, the question arises, are we able to improve the effectiveness and efficiency of tasks with no training examples through selective fine-tuning? In this paper, we explore statistical measures that approximate the divergence between domain representations as a means to estimate whether tuning using one task pair will exhibit performance benefits over tuning another. This estimation can then be used to reduce the number of task pairs that need to be tested by eliminating pairs that are unlikely to provide benefits. Through experimentation over 58 tasks and over 6,600 task pair combinations, we demonstrate that statistical measures can distinguish effective task pairs, and the resulting estimates can reduce end-to-end runtime by up to 40%.
从预先训练的神经语言模型中转移学习模式已被证明可以显著提高各种基于语言的任务的有效性,同时,如果中间任务与目标任务充分相关,则可以进一步调整中间任务,以提供额外的性能优势。然而,如何识别相关任务是一个悬而未决的问题,强力搜索有效的任务组合的成本高得令人望而却步。因此,问题来了,我们是否能够通过选择性微调,在没有训练实例的情况下提高任务的有效性和效率?在本文中,我们探索了近似域表示之间差异的统计度量,以此来估计使用一个任务对进行调优是否会比使用另一个任务组表现出性能优势。然后,通过消除不太可能带来好处的任务对,可以使用这种估计来减少需要测试的任务对的数量。通过对58个任务和6600多个任务对组合的实验,我们证明了统计测量可以区分有效的任务对,由此产生的估计可以将端到端运行时间减少40%。
{"title":"Divergence-Based Domain Transferability for Zero-Shot Classification","authors":"Alexander Pugantsov, R. McCreadie","doi":"10.48550/arXiv.2302.05735","DOIUrl":"https://doi.org/10.48550/arXiv.2302.05735","url":null,"abstract":"Transferring learned patterns from pretrained neural language models has been shown to significantly improve effectiveness across a variety of language-based tasks, meanwhile further tuning on intermediate tasks has been demonstrated to provide additional performance benefits, provided the intermediate task is sufficiently related to the target task. However, how to identify related tasks is an open problem, and brute-force searching effective task combinations is prohibitively expensive. Hence, the question arises, are we able to improve the effectiveness and efficiency of tasks with no training examples through selective fine-tuning? In this paper, we explore statistical measures that approximate the divergence between domain representations as a means to estimate whether tuning using one task pair will exhibit performance benefits over tuning another. This estimation can then be used to reduce the number of task pairs that need to be tested by eliminating pairs that are unlikely to provide benefits. Through experimentation over 58 tasks and over 6,600 task pair combinations, we demonstrate that statistical measures can distinguish effective task pairs, and the resulting estimates can reduce end-to-end runtime by up to 40%.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"1604-1609"},"PeriodicalIF":0.0,"publicationDate":"2023-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44795146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NapSS: Paragraph-level Medical Text Simplification via Narrative Prompting and Sentence-matching Summarization 通过叙述提示和句子匹配摘要的段落级医学文本简化
Pub Date : 2023-02-11 DOI: 10.48550/arXiv.2302.05574
Junru Lu, Jiazheng Li, Byron C. Wallace, Yulan He, Gabriele Pergola
Accessing medical literature is difficult for laypeople as the content is written for specialists and contains medical jargon. Automated text simplification methods offer a potential means to address this issue. In this work, we propose a summarize-then-simplify two-stage strategy, which we call NapSS, identifying the relevant content to simplify while ensuring that the original narrative flow is preserved. In this approach, we first generate reference summaries via sentence matching between the original and the simplified abstracts. These summaries are then used to train an extractive summarizer, learning the most relevant content to be simplified. Then, to ensure the narrative consistency of the simplified text, we synthesize auxiliary narrative prompts combining key phrases derived from the syntactical analyses of the original text. Our model achieves results significantly better than the seq2seq baseline on an English medical corpus, yielding 3%~4% absolute improvements in terms of lexical similarity, and providing a further 1.1% improvement of SARI score when combined with the baseline. We also highlight shortcomings of existing evaluation methods, and introduce new metrics that take into account both lexical and high-level semantic similarity. A human evaluation conducted on a random sample of the test set further establishes the effectiveness of the proposed approach. Codes and models are released here: https://github.com/LuJunru/NapSS.
非专业人士很难访问医学文献,因为这些内容是为专家撰写的,并且包含医学术语。自动化文本简化方法为解决这一问题提供了一种潜在的手段。在这项工作中,我们提出了一种总结然后简化的两阶段策略,我们称之为NapSS,确定要简化的相关内容,同时确保原始叙事流得到保留。在这种方法中,我们首先通过原始摘要和简化摘要之间的句子匹配来生成参考摘要。然后,这些摘要被用来训练提取式总结者,学习要简化的最相关的内容。然后,为了确保简化文本的叙事一致性,我们结合对原文句法分析得出的关键短语,合成辅助叙事提示。我们的模型在英语医学语料库中取得的结果明显好于seq2seq基线,在词汇相似性方面产生了3%~4%的绝对改善,与基线相结合时,严重急性呼吸系统综合征得分进一步提高了1.1%。我们还强调了现有评估方法的不足,并引入了同时考虑词汇和高级语义相似性的新指标。对测试集的随机样本进行的人类评估进一步证明了所提出方法的有效性。此处发布代码和型号:https://github.com/LuJunru/NapSS.
{"title":"NapSS: Paragraph-level Medical Text Simplification via Narrative Prompting and Sentence-matching Summarization","authors":"Junru Lu, Jiazheng Li, Byron C. Wallace, Yulan He, Gabriele Pergola","doi":"10.48550/arXiv.2302.05574","DOIUrl":"https://doi.org/10.48550/arXiv.2302.05574","url":null,"abstract":"Accessing medical literature is difficult for laypeople as the content is written for specialists and contains medical jargon. Automated text simplification methods offer a potential means to address this issue. In this work, we propose a summarize-then-simplify two-stage strategy, which we call NapSS, identifying the relevant content to simplify while ensuring that the original narrative flow is preserved. In this approach, we first generate reference summaries via sentence matching between the original and the simplified abstracts. These summaries are then used to train an extractive summarizer, learning the most relevant content to be simplified. Then, to ensure the narrative consistency of the simplified text, we synthesize auxiliary narrative prompts combining key phrases derived from the syntactical analyses of the original text. Our model achieves results significantly better than the seq2seq baseline on an English medical corpus, yielding 3%~4% absolute improvements in terms of lexical similarity, and providing a further 1.1% improvement of SARI score when combined with the baseline. We also highlight shortcomings of existing evaluation methods, and introduce new metrics that take into account both lexical and high-level semantic similarity. A human evaluation conducted on a random sample of the test set further establishes the effectiveness of the proposed approach. Codes and models are released here: https://github.com/LuJunru/NapSS.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"1049-1061"},"PeriodicalIF":0.0,"publicationDate":"2023-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48370804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Predicting Desirable Revisions of Evidence and Reasoning in Argumentative Writing 议论文写作中证据与推理的预期修正
Pub Date : 2023-02-10 DOI: 10.48550/arXiv.2302.05039
T. Afrin, D. Litman
We develop models to classify desirable evidence and desirable reasoning revisions in student argumentative writing. We explore two ways to improve classifier performance – using the essay context of the revision, and using the feedback students received before the revision. We perform both intrinsic and extrinsic evaluation for each of our models and report a qualitative analysis. Our results show that while a model using feedback information improves over a baseline model, models utilizing context - either alone or with feedback - are the most successful in identifying desirable revisions.
我们开发了模型来对学生议论文中的合意证据和合意推理修正进行分类。我们探索了两种提高分类器性能的方法——使用复习的文章上下文和使用学生在复习前收到的反馈。我们对每个模型进行内在和外在评估,并报告定性分析。我们的结果表明,虽然使用反馈信息的模型比基线模型有所改进,但使用上下文的模型——无论是单独的还是有反馈的——在确定理想的修订方面是最成功的。
{"title":"Predicting Desirable Revisions of Evidence and Reasoning in Argumentative Writing","authors":"T. Afrin, D. Litman","doi":"10.48550/arXiv.2302.05039","DOIUrl":"https://doi.org/10.48550/arXiv.2302.05039","url":null,"abstract":"We develop models to classify desirable evidence and desirable reasoning revisions in student argumentative writing. We explore two ways to improve classifier performance – using the essay context of the revision, and using the feedback students received before the revision. We perform both intrinsic and extrinsic evaluation for each of our models and report a qualitative analysis. Our results show that while a model using feedback information improves over a baseline model, models utilizing context - either alone or with feedback - are the most successful in identifying desirable revisions.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"2505-2516"},"PeriodicalIF":0.0,"publicationDate":"2023-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44902704","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Global Constraints with Prompting for Zero-Shot Event Argument Classification 带有提示的零射击事件参数分类的全局约束
Pub Date : 2023-02-09 DOI: 10.48550/arXiv.2302.04459
Zizheng Lin, Hongming Zhang, Yangqiu Song
Determining the role of event arguments is a crucial subtask of event extraction. Most previous supervised models leverage costly annotations, which is not practical for open-domain applications. In this work, we propose to use global constraints with prompting to effectively tackles event argument classification without any annotation and task-specific training. Specifically, given an event and its associated passage, the model first creates several new passages by prefix prompts and cloze prompts, where prefix prompts indicate event type and trigger span, and cloze prompts connect each candidate role with the target argument span. Then, a pre-trained language model scores the new passages, making the initial prediction. Our novel prompt templates can easily adapt to all events and argument types without manual effort. Next, the model regularizes the prediction by global constraints exploiting cross-task, cross-argument, and cross-event relations. Extensive experiments demonstrate our model’s effectiveness: it outperforms the best zero-shot baselines by 12.5% and 10.9% F1 on ACE and ERE with given argument spans and by 4.3% and 3.3% F1, respectively, without given argument spans. We have made our code publicly available.
确定事件参数的作用是事件提取的一个关键子任务。以前的大多数监督模型都利用了昂贵的注释,这对于开放域应用程序来说是不实用的。在这项工作中,我们建议使用带有提示的全局约束,在没有任何注释和任务特定训练的情况下有效地处理事件自变量分类。具体来说,给定一个事件及其相关段落,模型首先通过前缀提示和完形填空提示创建几个新段落,其中前缀提示指示事件类型和触发跨度,完形填空提醒将每个候选角色与目标论点跨度连接起来。然后,一个预先训练的语言模型对新段落进行评分,进行初步预测。我们新颖的提示模板可以轻松地适应所有事件和参数类型,而无需手动操作。接下来,该模型通过利用跨任务、跨自变量和跨事件关系的全局约束来正则化预测。大量实验证明了我们的模型的有效性:在给定变元跨度的情况下,它在ACE和ERE上分别比最佳零样本基线好12.5%和10.9%F1,在没有给定变元间距的情况下分别好4.3%和3.3%F1。我们已经公开了我们的代码。
{"title":"Global Constraints with Prompting for Zero-Shot Event Argument Classification","authors":"Zizheng Lin, Hongming Zhang, Yangqiu Song","doi":"10.48550/arXiv.2302.04459","DOIUrl":"https://doi.org/10.48550/arXiv.2302.04459","url":null,"abstract":"Determining the role of event arguments is a crucial subtask of event extraction. Most previous supervised models leverage costly annotations, which is not practical for open-domain applications. In this work, we propose to use global constraints with prompting to effectively tackles event argument classification without any annotation and task-specific training. Specifically, given an event and its associated passage, the model first creates several new passages by prefix prompts and cloze prompts, where prefix prompts indicate event type and trigger span, and cloze prompts connect each candidate role with the target argument span. Then, a pre-trained language model scores the new passages, making the initial prediction. Our novel prompt templates can easily adapt to all events and argument types without manual effort. Next, the model regularizes the prediction by global constraints exploiting cross-task, cross-argument, and cross-event relations. Extensive experiments demonstrate our model’s effectiveness: it outperforms the best zero-shot baselines by 12.5% and 10.9% F1 on ACE and ERE with given argument spans and by 4.3% and 3.3% F1, respectively, without given argument spans. We have made our code publicly available.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"2482-2493"},"PeriodicalIF":0.0,"publicationDate":"2023-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48698839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Large-Scale Multilingual Study of Visual Constraints on Linguistic Selection of Descriptions 描述语言选择的视觉约束的大规模多语言研究
Pub Date : 2023-02-09 DOI: 10.48550/arXiv.2302.04811
Uri Berger, Lea Frermann, G. Stanovsky, Omri Abend
We present a large, multilingual study into how vision constrains linguistic choice, covering four languages and five linguistic properties, such as verb transitivity or use of numerals. We propose a novel method that leverages existing corpora of images with captions written by native speakers, and apply it to nine corpora, comprising 600k images and 3M captions. We study the relation between visual input and linguistic choices by training classifiers to predict the probability of expressing a property from raw images, and find evidence supporting the claim that linguistic properties are constrained by visual context across languages. We complement this investigation with a corpus study, taking the test case of numerals. Specifically, we use existing annotations (number or type of objects) to investigate the effect of different visual conditions on the use of numeral expressions in captions, and show that similar patterns emerge across languages.Our methods and findings both confirm and extend existing research in the cognitive literature. We additionally discuss possible applications for language generation.
我们对视觉如何限制语言选择进行了一项大规模的多语言研究,涵盖了四种语言和五种语言特性,如动词及物性或数字的使用。我们提出了一种新的方法,利用现有的带有母语者写的字幕的图像语料库,并将其应用于9个语料库,包括600k个图像和3M个字幕。我们通过训练分类器来预测从原始图像中表达属性的概率,研究了视觉输入和语言选择之间的关系,并找到了支持语言属性受语言视觉上下文约束的证据。我们用语料库研究来补充这一调查,以数字为测试案例。具体来说,我们使用现有的注释(对象的数量或类型)来研究不同视觉条件对字幕中数字表达的影响,并表明不同语言之间会出现类似的模式。我们的方法和发现证实并扩展了认知文献中现有的研究。我们还讨论了语言生成的可能应用。
{"title":"A Large-Scale Multilingual Study of Visual Constraints on Linguistic Selection of Descriptions","authors":"Uri Berger, Lea Frermann, G. Stanovsky, Omri Abend","doi":"10.48550/arXiv.2302.04811","DOIUrl":"https://doi.org/10.48550/arXiv.2302.04811","url":null,"abstract":"We present a large, multilingual study into how vision constrains linguistic choice, covering four languages and five linguistic properties, such as verb transitivity or use of numerals. We propose a novel method that leverages existing corpora of images with captions written by native speakers, and apply it to nine corpora, comprising 600k images and 3M captions. We study the relation between visual input and linguistic choices by training classifiers to predict the probability of expressing a property from raw images, and find evidence supporting the claim that linguistic properties are constrained by visual context across languages. We complement this investigation with a corpus study, taking the test case of numerals. Specifically, we use existing annotations (number or type of objects) to investigate the effect of different visual conditions on the use of numeral expressions in captions, and show that similar patterns emerge across languages.Our methods and findings both confirm and extend existing research in the cognitive literature. We additionally discuss possible applications for language generation.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"2240-2254"},"PeriodicalIF":0.0,"publicationDate":"2023-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46634111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Revisiting Offline Compression: Going Beyond Factorization-based Methods for Transformer Language Models 重新审视离线压缩:超越基于分解的转换语言模型方法
Pub Date : 2023-02-08 DOI: 10.48550/arXiv.2302.04045
Mohammadreza Banaei, Klaudia Bałazy, A. Kasymov, R. Lebret, J. Tabor, K. Aberer
Recent transformer language models achieve outstanding results in many natural language processing (NLP) tasks. However, their enormous size often makes them impractical on memory-constrained devices, requiring practitioners to compress them to smaller networks. In this paper, we explore offline compression methods, meaning computationally-cheap approaches that do not require further fine-tuning of the compressed model. We challenge the classical matrix factorization methods by proposing a novel, better-performing autoencoder-based framework. We perform a comprehensive ablation study of our approach, examining its different aspects over a diverse set of evaluation settings. Moreover, we show that enabling collaboration between modules across layers by compressing certain modules together positively impacts the final model performance. Experiments on various NLP tasks demonstrate that our approach significantly outperforms commonly used factorization-based offline compression methods.
最近的转换语言模型在许多自然语言处理(NLP)任务中取得了出色的成绩。然而,它们的巨大尺寸往往使它们在内存受限的设备上不切实际,需要从业者将它们压缩到更小的网络上。在本文中,我们探索了离线压缩方法,即不需要进一步微调压缩模型的计算廉价方法。我们通过提出一种新的、性能更好的基于自编码器的框架来挑战经典的矩阵分解方法。我们对我们的方法进行了全面的消融研究,在不同的评估设置中检查了它的不同方面。此外,我们表明,通过将某些模块压缩在一起来实现跨层模块之间的协作对最终模型性能有积极影响。在各种NLP任务上的实验表明,我们的方法明显优于常用的基于因子分解的离线压缩方法。
{"title":"Revisiting Offline Compression: Going Beyond Factorization-based Methods for Transformer Language Models","authors":"Mohammadreza Banaei, Klaudia Bałazy, A. Kasymov, R. Lebret, J. Tabor, K. Aberer","doi":"10.48550/arXiv.2302.04045","DOIUrl":"https://doi.org/10.48550/arXiv.2302.04045","url":null,"abstract":"Recent transformer language models achieve outstanding results in many natural language processing (NLP) tasks. However, their enormous size often makes them impractical on memory-constrained devices, requiring practitioners to compress them to smaller networks. In this paper, we explore offline compression methods, meaning computationally-cheap approaches that do not require further fine-tuning of the compressed model. We challenge the classical matrix factorization methods by proposing a novel, better-performing autoencoder-based framework. We perform a comprehensive ablation study of our approach, examining its different aspects over a diverse set of evaluation settings. Moreover, we show that enabling collaboration between modules across layers by compressing certain modules together positively impacts the final model performance. Experiments on various NLP tasks demonstrate that our approach significantly outperforms commonly used factorization-based offline compression methods.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"23 1","pages":"1743-1760"},"PeriodicalIF":0.0,"publicationDate":"2023-02-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91126186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mining Effective Features Using Quantum Entropy for Humor Recognition 利用量子熵挖掘幽默识别的有效特征
Pub Date : 2023-02-07 DOI: 10.48550/arXiv.2302.03716
Y. Liu, Yuexian Hou
Humor recognition has been extensively studied with different methods in the past years. However, existing studies on humor recognition do not understand the mechanisms that generate humor. In this paper, inspired by the incongruity theory, any joke can be divided into two components (the setup and the punchline). Both components have multiple possible semantics, and there is an incongruous relationship between them. We use density matrices to represent the semantic uncertainty of the setup and the punchline, respectively, and design QE-Uncertainty and QE-Incongruity with the help of quantum entropy as features for humor recognition. The experimental results on the SemEval2021 Task 7 dataset show that the proposed features are more effective than the baselines for recognizing humorous and non-humorous texts.
幽默识别在过去的几年里已经用不同的方法进行了广泛的研究。然而,现有关于幽默识别的研究并不了解幽默产生的机制。在本文中,受不协调理论的启发,任何笑话都可以分为两个部分(设置和笑点)。这两个组件都有多种可能的语义,并且它们之间存在不协调的关系。我们使用密度矩阵分别表示设置和笑点的语义不确定性,并在量子熵的帮助下设计QE不确定性和QE不一致性作为幽默识别的特征。在SemEval2021 Task 7数据集上的实验结果表明,在识别幽默和非幽默文本方面,所提出的特征比基线更有效。
{"title":"Mining Effective Features Using Quantum Entropy for Humor Recognition","authors":"Y. Liu, Yuexian Hou","doi":"10.48550/arXiv.2302.03716","DOIUrl":"https://doi.org/10.48550/arXiv.2302.03716","url":null,"abstract":"Humor recognition has been extensively studied with different methods in the past years. However, existing studies on humor recognition do not understand the mechanisms that generate humor. In this paper, inspired by the incongruity theory, any joke can be divided into two components (the setup and the punchline). Both components have multiple possible semantics, and there is an incongruous relationship between them. We use density matrices to represent the semantic uncertainty of the setup and the punchline, respectively, and design QE-Uncertainty and QE-Incongruity with the help of quantum entropy as features for humor recognition. The experimental results on the SemEval2021 Task 7 dataset show that the proposed features are more effective than the baselines for recognizing humorous and non-humorous texts.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"2003-2008"},"PeriodicalIF":0.0,"publicationDate":"2023-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47399107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Entity-Aware Dual Co-Attention Network for Fake News Detection 基于实体感知的假新闻检测双共注意网络
Pub Date : 2023-02-07 DOI: 10.48550/arXiv.2302.03475
Sin-Han Yang, Chung-Chi Chen, Hen-Hsen Huang, Hsin-Hsi Chen
Fake news and misinformation spread rapidly on the Internet. How to identify it and how to interpret the identification results have become important issues. In this paper, we propose a Dual Co-Attention Network (Dual-CAN) for fake news detection, which takes news content, social media replies, and external knowledge into consideration. Our experimental results support that the proposed Dual-CAN outperforms current representative models in two benchmark datasets. We further make in-depth discussions by comparing how models work in both datasets with empirical analysis of attention weights.
假新闻和错误信息在互联网上迅速传播。如何识别它以及如何解释识别结果已成为重要问题。在本文中,我们提出了一种用于假新闻检测的双共同注意网络(Dual CAN),该网络考虑了新闻内容、社交媒体回复和外部知识。我们的实验结果支持所提出的双CAN在两个基准数据集中优于当前的代表性模型。我们通过比较模型在两个数据集中的工作方式和注意力权重的实证分析,进一步进行了深入的讨论。
{"title":"Entity-Aware Dual Co-Attention Network for Fake News Detection","authors":"Sin-Han Yang, Chung-Chi Chen, Hen-Hsen Huang, Hsin-Hsi Chen","doi":"10.48550/arXiv.2302.03475","DOIUrl":"https://doi.org/10.48550/arXiv.2302.03475","url":null,"abstract":"Fake news and misinformation spread rapidly on the Internet. How to identify it and how to interpret the identification results have become important issues. In this paper, we propose a Dual Co-Attention Network (Dual-CAN) for fake news detection, which takes news content, social media replies, and external knowledge into consideration. Our experimental results support that the proposed Dual-CAN outperforms current representative models in two benchmark datasets. We further make in-depth discussions by comparing how models work in both datasets with empirical analysis of attention weights.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"106-113"},"PeriodicalIF":0.0,"publicationDate":"2023-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45783595","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Findings (Sydney (N.S.W.)
全部 Geobiology Carbon Balance Manage. Geosci. J. Ocean and Coastal Research Global Biogeochem. Cycles B SOC GEOL MEX ARCHAEOMETRY ACTA GEOL POL OCEAN SCI J QUATERNAIRE J. Atmos. Sol. Terr. Phys. GEOHERITAGE IDOJARAS Environ. Mol. Mutagen. ARCT ANTARCT ALP RES Chem. Ecol. TERR ATMOS OCEAN SCI Precambrian Res. GROUNDWATER J. Clim. INDIAN J GEO-MAR SCI GEOCHEM GEOPHY GEOSY Surv. Geophys. Terra Nova 2013 IEEE International Symposium on Hardware-Oriented Security and Trust (HOST) ENVIRON HEALTH-GLOB Atmos. Res. Ann. Glaciol. Environmental Claims Journal Enzyme Research Surv. Rev. J. South Am. Earth Sci. NEUES JB MINER ABH Paleontol. Res. Prog. Oceanogr. Adv. Atmos. Sci. 2013 21st IEEE International Requirements Engineering Conference (RE) J APPL METEOROL CLIM ENVIRONMENT Acta Palaeontol. Polonica J. Environ. Eng. Geophys. OCEANOLOGY+ Org. Geochem. J. Seismic Explor. ENTROPY-SWITZ 2008 International Conference on Electronic Packaging Technology & High Density Packaging ITAL J REMOTE SENS Acta Montan. Slovaca Space Weather Eur. J. Control Geol. Ore Deposits Int. J. Earth Sci. Aust. J. Earth Sci. Am. J. Sci. ACTA PETROL SIN GEOLOGY Atmos. Chem. Phys. Asia-Pac. J. Atmos. Sci. ACTA GEOL SIN-ENGL Am. Mineral. Archaeol. Anthropol. Sci. IZV-PHYS SOLID EART+ Acta Geophys. 2009 12th International Symposium on Design and Diagnostics of Electronic Circuits & Systems Geostand. Geoanal. Res. Acta Oceanolog. Sin. Atmos. Meas. Tech. BIOGEOSCIENCES J. Atmos. Oceanic Technol. Basin Res. J. Atmos. Chem. Adv. Meteorol. Appl. Geochem. Hydrol. Earth Syst. Sci. Acta Geochimica Int. J. Climatol. AAPG Bull. Am. J. Phys. Anthropol. Miner. Deposita Big Earth Data Yan Ke Xue Bao (Hong Kong) ATMOSPHERE-BASEL J. Hydrol. Can. J. Phys. Nat. Geosci. THALASSAS Nat. Hazards Earth Syst. Sci. Ore Geol. Rev. Appl. Clay Sci. Nat. Clim. Change Mon. Weather Rev. 2013 IEEE International Solid-State Circuits Conference Digest of Technical Papers J. Meteorolog. Res. Ocean Dyn. Contrib. Mineral. Petrol. J. Geog. Sci. ArcheoSci.-Rev. Archeom. Int. J. Geomech. Annu. Rev. Earth Planet. Sci. Resour. Geol.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1