首页 > 最新文献

AACL Bioflux最新文献

英文 中文
Every word counts: A multilingual analysis of individual human alignment with model attention 每个单词计数:一个多语言的分析,个人与模型的注意力对齐
Q3 Environmental Science Pub Date : 2022-10-05 DOI: 10.48550/arXiv.2210.04963
Stephanie Brandl, Nora Hollenstein
Human fixation patterns have been shown to correlate strongly with Transformer-based attention. Those correlation analyses are usually carried out without taking into account individual differences between participants and are mostly done on monolingual datasets making it difficult to generalise findings. In this paper, we analyse eye-tracking data from speakers of 13 different languages reading both in their native language (L1) and in English as language learners (L2). We find considerable differences between languages but also that individual reading behaviour such as skipping rate, total reading time and vocabulary knowledge (LexTALE) influence the alignment between humans and models to an extent that should be considered in future studies.
人类固定模式已被证明与基于变形金刚的注意力密切相关。这些相关性分析通常在没有考虑参与者之间的个体差异的情况下进行,并且主要是在单语言数据集上进行的,因此很难概括发现。在本文中,我们分析了13种不同语言的使用者以母语(第一语言)和作为语言学习者的英语(第二语言)阅读时的眼动追踪数据。我们发现语言之间存在相当大的差异,但个人阅读行为(如跳过率、总阅读时间和词汇知识(LexTALE))也会在一定程度上影响人类和模型之间的一致性,这应该在未来的研究中得到考虑。
{"title":"Every word counts: A multilingual analysis of individual human alignment with model attention","authors":"Stephanie Brandl, Nora Hollenstein","doi":"10.48550/arXiv.2210.04963","DOIUrl":"https://doi.org/10.48550/arXiv.2210.04963","url":null,"abstract":"Human fixation patterns have been shown to correlate strongly with Transformer-based attention. Those correlation analyses are usually carried out without taking into account individual differences between participants and are mostly done on monolingual datasets making it difficult to generalise findings. In this paper, we analyse eye-tracking data from speakers of 13 different languages reading both in their native language (L1) and in English as language learners (L2). We find considerable differences between languages but also that individual reading behaviour such as skipping rate, total reading time and vocabulary knowledge (LexTALE) influence the alignment between humans and models to an extent that should be considered in future studies.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"06 1","pages":"72-77"},"PeriodicalIF":0.0,"publicationDate":"2022-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86148826","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
An Embarrassingly Simple Approach for Intellectual Property Rights Protection on Recurrent Neural Networks 递归神经网络知识产权保护的简单方法
Q3 Environmental Science Pub Date : 2022-10-03 DOI: 10.48550/arXiv.2210.00743
Zhi Qin Tan, H. P. Wong, Chee Seng Chan
Capitalise on deep learning models, offering Natural Language Processing (NLP) solutions as a part of the Machine Learning as a Service (MLaaS) has generated handsome revenues. At the same time, it is known that the creation of these lucrative deep models is non-trivial. Therefore, protecting these inventions’ intellectual property rights (IPR) from being abused, stolen and plagiarized is vital. This paper proposes a practical approach for the IPR protection on recurrent neural networks (RNN) without all the bells and whistles of existing IPR solutions. Particularly, we introduce the Gatekeeper concept that resembles the recurrent nature in RNN architecture to embed keys. Also, we design the model training scheme in a way such that the protected RNN model will retain its original performance iff a genuine key is presented. Extensive experiments showed that our protection scheme is robust and effective against ambiguity and removal attacks in both white-box and black-box protection schemes on different RNN variants. Code is available at https://github.com/zhiqin1998/RecurrentIPR.
利用深度学习模型,提供自然语言处理(NLP)解决方案,作为机器学习即服务(MLaaS)的一部分,已经产生了可观的收入。与此同时,众所周知,这些有利可图的深度模型的创建是不平凡的。因此,保护这些发明的知识产权不被滥用、窃取和剽窃是至关重要的。本文提出了一种实用的递归神经网络(RNN)知识产权保护方法,该方法不需要现有知识产权解决方案的所有附加功能。特别地,我们引入了类似于RNN架构中循环性质的看门人概念来嵌入密钥。此外,我们设计了模型训练方案,使受保护的RNN模型在提供真实密钥时保持其原始性能。大量的实验表明,我们的保护方案对不同RNN变体的白盒和黑盒保护方案都具有鲁棒性和有效性。代码可从https://github.com/zhiqin1998/RecurrentIPR获得。
{"title":"An Embarrassingly Simple Approach for Intellectual Property Rights Protection on Recurrent Neural Networks","authors":"Zhi Qin Tan, H. P. Wong, Chee Seng Chan","doi":"10.48550/arXiv.2210.00743","DOIUrl":"https://doi.org/10.48550/arXiv.2210.00743","url":null,"abstract":"Capitalise on deep learning models, offering Natural Language Processing (NLP) solutions as a part of the Machine Learning as a Service (MLaaS) has generated handsome revenues. At the same time, it is known that the creation of these lucrative deep models is non-trivial. Therefore, protecting these inventions’ intellectual property rights (IPR) from being abused, stolen and plagiarized is vital. This paper proposes a practical approach for the IPR protection on recurrent neural networks (RNN) without all the bells and whistles of existing IPR solutions. Particularly, we introduce the Gatekeeper concept that resembles the recurrent nature in RNN architecture to embed keys. Also, we design the model training scheme in a way such that the protected RNN model will retain its original performance iff a genuine key is presented. Extensive experiments showed that our protection scheme is robust and effective against ambiguity and removal attacks in both white-box and black-box protection schemes on different RNN variants. Code is available at https://github.com/zhiqin1998/RecurrentIPR.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"42 1","pages":"93-105"},"PeriodicalIF":0.0,"publicationDate":"2022-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89262045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Fine-grained Contrastive Learning for Definition Generation 定义生成的细粒度对比学习
Q3 Environmental Science Pub Date : 2022-10-02 DOI: 10.48550/arXiv.2210.00543
Hengyuan Zhang, Dawei Li, Shiping Yang, Yanran Li
Recently, pre-trained transformer-based models have achieved great success in the task of definition generation (DG). However, previous encoder-decoder models lack effective representation learning to contain full semantic components of the given word, which leads to generating under-specific definitions. To address this problem, we propose a novel contrastive learning method, encouraging the model to capture more detailed semantic representations from the definition sequence encoding. According to both automatic and manual evaluation, the experimental results on three mainstream benchmarks demonstrate that the proposed method could generate more specific and high-quality definitions compared with several state-of-the-art models.
近年来,基于预训练变压器的模型在定义生成(DG)任务中取得了巨大成功。然而,以前的编码器-解码器模型缺乏有效的表示学习来包含给定单词的完整语义组件,这导致生成不特定的定义。为了解决这个问题,我们提出了一种新的对比学习方法,鼓励模型从定义序列编码中捕获更详细的语义表示。根据自动和手动评估,在三个主流基准上的实验结果表明,与几种最先进的模型相比,该方法可以生成更具体和高质量的定义。
{"title":"Fine-grained Contrastive Learning for Definition Generation","authors":"Hengyuan Zhang, Dawei Li, Shiping Yang, Yanran Li","doi":"10.48550/arXiv.2210.00543","DOIUrl":"https://doi.org/10.48550/arXiv.2210.00543","url":null,"abstract":"Recently, pre-trained transformer-based models have achieved great success in the task of definition generation (DG). However, previous encoder-decoder models lack effective representation learning to contain full semantic components of the given word, which leads to generating under-specific definitions. To address this problem, we propose a novel contrastive learning method, encouraging the model to capture more detailed semantic representations from the definition sequence encoding. According to both automatic and manual evaluation, the experimental results on three mainstream benchmarks demonstrate that the proposed method could generate more specific and high-quality definitions compared with several state-of-the-art models.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"112 1","pages":"1001-1012"},"PeriodicalIF":0.0,"publicationDate":"2022-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82465931","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Risk-graded Safety for Handling Medical Queries in Conversational AI 会话AI中处理医疗查询的风险分级安全性
Q3 Environmental Science Pub Date : 2022-10-02 DOI: 10.48550/arXiv.2210.00572
Gavin Abercrombie, Verena Rieser
Conversational AI systems can engage in unsafe behaviour when handling users’ medical queries that may have severe consequences and could even lead to deaths. Systems therefore need to be capable of both recognising the seriousness of medical inputs and producing responses with appropriate levels of risk. We create a corpus of human written English language medical queries and the responses of different types of systems. We label these with both crowdsourced and expert annotations. While individual crowdworkers may be unreliable at grading the seriousness of the prompts, their aggregated labels tend to agree with professional opinion to a greater extent on identifying the medical queries and recognising the risk types posed by the responses. Results of classification experiments suggest that, while these tasks can be automated, caution should be exercised, as errors can potentially be very serious.
会话式人工智能系统在处理用户的医疗查询时可能会采取不安全的行为,这可能会产生严重的后果,甚至可能导致死亡。因此,系统需要既能认识到医疗投入的严重性,又能以适当的风险水平作出反应。我们创建了一个人类书面英语医学查询和不同类型系统响应的语料库。我们用众包注释和专家注释来标记它们。虽然个别众包工作者在对提示的严重性进行分级方面可能不可靠,但他们的综合标签在识别医疗问题和识别回答所带来的风险类型方面往往更符合专业意见。分类实验的结果表明,虽然这些任务可以自动化,但应该谨慎行事,因为错误可能非常严重。
{"title":"Risk-graded Safety for Handling Medical Queries in Conversational AI","authors":"Gavin Abercrombie, Verena Rieser","doi":"10.48550/arXiv.2210.00572","DOIUrl":"https://doi.org/10.48550/arXiv.2210.00572","url":null,"abstract":"Conversational AI systems can engage in unsafe behaviour when handling users’ medical queries that may have severe consequences and could even lead to deaths. Systems therefore need to be capable of both recognising the seriousness of medical inputs and producing responses with appropriate levels of risk. We create a corpus of human written English language medical queries and the responses of different types of systems. We label these with both crowdsourced and expert annotations. While individual crowdworkers may be unreliable at grading the seriousness of the prompts, their aggregated labels tend to agree with professional opinion to a greater extent on identifying the medical queries and recognising the risk types posed by the responses. Results of classification experiments suggest that, while these tasks can be automated, caution should be exercised, as errors can potentially be very serious.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"100 1","pages":"234-243"},"PeriodicalIF":0.0,"publicationDate":"2022-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81023036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Decade of Knowledge Graphs in Natural Language Processing: A Survey 十年来自然语言处理中的知识图谱:综述
Q3 Environmental Science Pub Date : 2022-09-30 DOI: 10.48550/arXiv.2210.00105
Phillip Schneider, Tim Schopf, Juraj Vladika, Mikhail Galkin, E. Simperl, F. Matthes
In pace with developments in the research field of artificial intelligence, knowledge graphs (KGs) have attracted a surge of interest from both academia and industry. As a representation of semantic relations between entities, KGs have proven to be particularly relevant for natural language processing (NLP), experiencing a rapid spread and wide adoption within recent years. Given the increasing amount of research work in this area, several KG-related approaches have been surveyed in the NLP research community. However, a comprehensive study that categorizes established topics and reviews the maturity of individual research streams remains absent to this day. Contributing to closing this gap, we systematically analyzed 507 papers from the literature on KGs in NLP. Our survey encompasses a multifaceted review of tasks, research types, and contributions. As a result, we present a structured overview of the research landscape, provide a taxonomy of tasks, summarize our findings, and highlight directions for future work.
随着人工智能研究领域的发展,知识图谱(knowledge graphs, KGs)引起了学术界和工业界的极大兴趣。作为实体之间语义关系的表示,KGs已被证明与自然语言处理(NLP)特别相关,近年来经历了快速传播和广泛采用。鉴于这一领域的研究工作越来越多,NLP研究界已经调查了几种与kg相关的方法。然而,迄今为止,对已建立的主题进行分类并审查个别研究流成熟度的综合研究仍然缺乏。为了缩小这一差距,我们系统地分析了NLP中KGs文献中的507篇论文。我们的调查涵盖了任务、研究类型和贡献的多方面审查。因此,我们对研究前景进行了结构化的概述,提供了任务分类,总结了我们的发现,并强调了未来工作的方向。
{"title":"A Decade of Knowledge Graphs in Natural Language Processing: A Survey","authors":"Phillip Schneider, Tim Schopf, Juraj Vladika, Mikhail Galkin, E. Simperl, F. Matthes","doi":"10.48550/arXiv.2210.00105","DOIUrl":"https://doi.org/10.48550/arXiv.2210.00105","url":null,"abstract":"In pace with developments in the research field of artificial intelligence, knowledge graphs (KGs) have attracted a surge of interest from both academia and industry. As a representation of semantic relations between entities, KGs have proven to be particularly relevant for natural language processing (NLP), experiencing a rapid spread and wide adoption within recent years. Given the increasing amount of research work in this area, several KG-related approaches have been surveyed in the NLP research community. However, a comprehensive study that categorizes established topics and reviews the maturity of individual research streams remains absent to this day. Contributing to closing this gap, we systematically analyzed 507 papers from the literature on KGs in NLP. Our survey encompasses a multifaceted review of tasks, research types, and contributions. As a result, we present a structured overview of the research landscape, provide a taxonomy of tasks, summarize our findings, and highlight directions for future work.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"148 1","pages":"601-614"},"PeriodicalIF":0.0,"publicationDate":"2022-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77360821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
How to tackle an emerging topic? Combining strong and weak labels for Covid news NER 如何应对新出现的话题?结合Covid新闻NER的强标签和弱标签
Q3 Environmental Science Pub Date : 2022-09-29 DOI: 10.48550/arXiv.2209.15108
Aleksander Ficek, Fangyu Liu, Nigel Collier
Being able to train Named Entity Recognition (NER) models for emerging topics is crucial for many real-world applications especially in the medical domain where new topics are continuously evolving out of the scope of existing models and datasets. For a realistic evaluation setup, we introduce a novel COVID-19 news NER dataset (COVIDNEWS-NER) and release 3000 entries of hand annotated strongly labelled sentences and 13000 auto-generated weakly labelled sentences. Besides the dataset, we propose CONTROSTER, a recipe to strategically combine weak and strong labels in improving NER in an emerging topic through transfer learning. We show the effectiveness of CONTROSTER on COVIDNEWS-NER while providing analysis on combining weak and strong labels for training. Our key findings are: (1) Using weak data to formulate an initial backbone before tuning on strong data outperforms methods trained on only strong or weak data. (2) A combination of out-of-domain and in-domain weak label training is crucial and can overcome saturation when being training on weak labels from a single source.
能够训练新兴主题的命名实体识别(NER)模型对于许多实际应用至关重要,特别是在医学领域,新主题不断从现有模型和数据集的范围中发展出来。为了实现一个现实的评估设置,我们引入了一个新的COVID-19新闻NER数据集(covid - news -NER),并发布了3000条手工标注的强标记句子和13000条自动生成的弱标记句子。除了数据集,我们还提出了CONTROSTER,这是一种通过迁移学习战略性地结合弱标签和强标签来提高新兴主题NER的方法。我们展示了CONTROSTER对COVIDNEWS-NER的有效性,同时提供了结合弱标签和强标签进行训练的分析。我们的主要发现是:(1)在对强数据进行调优之前,使用弱数据来制定初始骨干,优于仅对强数据或弱数据进行训练的方法。(2)域外和域内弱标签训练的结合是至关重要的,可以克服单源弱标签训练时的饱和。
{"title":"How to tackle an emerging topic? Combining strong and weak labels for Covid news NER","authors":"Aleksander Ficek, Fangyu Liu, Nigel Collier","doi":"10.48550/arXiv.2209.15108","DOIUrl":"https://doi.org/10.48550/arXiv.2209.15108","url":null,"abstract":"Being able to train Named Entity Recognition (NER) models for emerging topics is crucial for many real-world applications especially in the medical domain where new topics are continuously evolving out of the scope of existing models and datasets. For a realistic evaluation setup, we introduce a novel COVID-19 news NER dataset (COVIDNEWS-NER) and release 3000 entries of hand annotated strongly labelled sentences and 13000 auto-generated weakly labelled sentences. Besides the dataset, we propose CONTROSTER, a recipe to strategically combine weak and strong labels in improving NER in an emerging topic through transfer learning. We show the effectiveness of CONTROSTER on COVIDNEWS-NER while providing analysis on combining weak and strong labels for training. Our key findings are: (1) Using weak data to formulate an initial backbone before tuning on strong data outperforms methods trained on only strong or weak data. (2) A combination of out-of-domain and in-domain weak label training is crucial and can overcome saturation when being training on weak labels from a single source.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"46 1","pages":"488-496"},"PeriodicalIF":0.0,"publicationDate":"2022-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87464846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Domain-aware Self-supervised Pre-training for Label-Efficient Meme Analysis 标签高效模因分析的领域感知自监督预训练
Q3 Environmental Science Pub Date : 2022-09-29 DOI: 10.48550/arXiv.2209.14667
Shivam Sharma, Mohd Khizir Siddiqui, Md. Shad Akhtar, Tanmoy Chakraborty
Existing self-supervised learning strategies are constrained to either a limited set of objectives or generic downstream tasks that predominantly target uni-modal applications. This has isolated progress for imperative multi-modal applications that are diverse in terms of complexity and domain-affinity, such as meme analysis. Here, we introduce two self-supervised pre-training methods, namely Ext-PIE-Net and MM-SimCLR that (i) employ off-the-shelf multi-modal hate-speech data during pre-training and (ii) perform self-supervised learning by incorporating multiple specialized pretext tasks, effectively catering to the required complex multi-modal representation learning for meme analysis. We experiment with different self-supervision strategies, including potential variants that could help learn rich cross-modality representations and evaluate using popular linear probing on the Hateful Memes task. The proposed solutions strongly compete with the fully supervised baseline via label-efficient training while distinctly outperforming them on all three tasks of the Memotion challenge with 0.18%, 23.64%, and 0.93% performance gain, respectively. Further, we demonstrate the generalizability of the proposed solutions by reporting competitive performance on the HarMeme task. Finally, we empirically establish the quality of the learned representations by analyzing task-specific learning, using fewer labeled training samples, and arguing that the complexity of the self-supervision strategy and downstream task at hand are correlated. Our efforts highlight the requirement of better multi-modal self-supervision methods involving specialized pretext tasks for efficient fine-tuning and generalizable performance.
现有的自监督学习策略要么局限于一组有限的目标,要么局限于主要针对单模态应用的一般下游任务。这隔离了命令式多模态应用程序的进展,这些应用程序在复杂性和领域亲和性方面各不相同,例如模因分析。本文介绍了ext -饼- net和MM-SimCLR两种自监督预训练方法(i)在预训练中使用现成的多模态仇恨言论数据;(ii)通过结合多个专门的借口任务进行自监督学习,有效地满足了模因分析所需的复杂多模态表征学习。我们尝试了不同的自我监督策略,包括可能有助于学习丰富的跨模态表征的潜在变体,并使用流行的线性探测对仇恨模因任务进行评估。提出的解决方案通过标签效率训练与完全监督基线进行激烈竞争,同时在Memotion挑战的所有三个任务上分别以0.18%,23.64%和0.93%的性能增益明显优于完全监督基线。此外,我们通过报告在HarMeme任务上的竞争表现来证明所提出解决方案的普遍性。最后,我们通过分析特定任务的学习,使用较少的标记训练样本,实证地建立了学习表征的质量,并认为自我监督策略的复杂性与手头的下游任务是相关的。我们的努力强调需要更好的多模态自我监督方法,包括专门的借口任务,以实现有效的微调和可推广的性能。
{"title":"Domain-aware Self-supervised Pre-training for Label-Efficient Meme Analysis","authors":"Shivam Sharma, Mohd Khizir Siddiqui, Md. Shad Akhtar, Tanmoy Chakraborty","doi":"10.48550/arXiv.2209.14667","DOIUrl":"https://doi.org/10.48550/arXiv.2209.14667","url":null,"abstract":"Existing self-supervised learning strategies are constrained to either a limited set of objectives or generic downstream tasks that predominantly target uni-modal applications. This has isolated progress for imperative multi-modal applications that are diverse in terms of complexity and domain-affinity, such as meme analysis. Here, we introduce two self-supervised pre-training methods, namely Ext-PIE-Net and MM-SimCLR that (i) employ off-the-shelf multi-modal hate-speech data during pre-training and (ii) perform self-supervised learning by incorporating multiple specialized pretext tasks, effectively catering to the required complex multi-modal representation learning for meme analysis. We experiment with different self-supervision strategies, including potential variants that could help learn rich cross-modality representations and evaluate using popular linear probing on the Hateful Memes task. The proposed solutions strongly compete with the fully supervised baseline via label-efficient training while distinctly outperforming them on all three tasks of the Memotion challenge with 0.18%, 23.64%, and 0.93% performance gain, respectively. Further, we demonstrate the generalizability of the proposed solutions by reporting competitive performance on the HarMeme task. Finally, we empirically establish the quality of the learned representations by analyzing task-specific learning, using fewer labeled training samples, and arguing that the complexity of the self-supervision strategy and downstream task at hand are correlated. Our efforts highlight the requirement of better multi-modal self-supervision methods involving specialized pretext tasks for efficient fine-tuning and generalizable performance.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"85 1","pages":"792-805"},"PeriodicalIF":0.0,"publicationDate":"2022-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80434682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Towards Simple and Efficient Task-Adaptive Pre-training for Text Classification 面向简单高效的任务自适应文本分类预训练
Q3 Environmental Science Pub Date : 2022-09-26 DOI: 10.48550/arXiv.2209.12943
Arnav Ladkat, Aamir Miyajiwala, Samiksha Jagadale, Rekha Kulkarni, Raviraj Joshi
Language models are pre-trained using large corpora of generic data like book corpus, com- mon crawl and Wikipedia, which is essential for the model to understand the linguistic characteristics of the language. New studies suggest using Domain Adaptive Pre-training (DAPT) and Task-Adaptive Pre-training (TAPT) as an intermediate step before the final finetuning task. This step helps cover the target domain vocabulary and improves the model performance on the downstream task. In this work, we study the impact of training only the embedding layer on the model’s performance during TAPT and task-specific finetuning. Based on our study, we propose a simple approach to make the in- termediate step of TAPT for BERT-based mod- els more efficient by performing selective pre-training of BERT layers. We show that training only the BERT embedding layer during TAPT is sufficient to adapt to the vocabulary of the target domain and achieve comparable performance. Our approach is computationally efficient, with 78% fewer parameters trained during TAPT. The proposed embedding layer finetuning approach can also be an efficient domain adaptation technique.
语言模型使用大量的通用数据(如图书语料库、common crawl和维基百科)进行预训练,这对于模型理解语言的语言特征至关重要。新的研究建议在最终调优任务之前使用域自适应预训练(DAPT)和任务自适应预训练(TAPT)作为中间步骤。这一步有助于覆盖目标领域词汇表,并提高下游任务的模型性能。在这项工作中,我们研究了在TAPT和特定任务微调期间只训练嵌入层对模型性能的影响。基于我们的研究,我们提出了一种简单的方法,通过对BERT层进行选择性预训练,使基于BERT的模型的中间步骤更有效。研究表明,在TAPT过程中只训练BERT嵌入层就足以适应目标领域的词汇表,并达到相当的性能。我们的方法计算效率很高,在TAPT期间训练的参数减少了78%。所提出的嵌入层微调方法也是一种有效的领域自适应技术。
{"title":"Towards Simple and Efficient Task-Adaptive Pre-training for Text Classification","authors":"Arnav Ladkat, Aamir Miyajiwala, Samiksha Jagadale, Rekha Kulkarni, Raviraj Joshi","doi":"10.48550/arXiv.2209.12943","DOIUrl":"https://doi.org/10.48550/arXiv.2209.12943","url":null,"abstract":"Language models are pre-trained using large corpora of generic data like book corpus, com- mon crawl and Wikipedia, which is essential for the model to understand the linguistic characteristics of the language. New studies suggest using Domain Adaptive Pre-training (DAPT) and Task-Adaptive Pre-training (TAPT) as an intermediate step before the final finetuning task. This step helps cover the target domain vocabulary and improves the model performance on the downstream task. In this work, we study the impact of training only the embedding layer on the model’s performance during TAPT and task-specific finetuning. Based on our study, we propose a simple approach to make the in- termediate step of TAPT for BERT-based mod- els more efficient by performing selective pre-training of BERT layers. We show that training only the BERT embedding layer during TAPT is sufficient to adapt to the vocabulary of the target domain and achieve comparable performance. Our approach is computationally efficient, with 78% fewer parameters trained during TAPT. The proposed embedding layer finetuning approach can also be an efficient domain adaptation technique.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"43 1","pages":"320-325"},"PeriodicalIF":0.0,"publicationDate":"2022-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82060798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Do ever larger octopi still amplify reporting biases? Evidence from judgments of typical colour 更大的章鱼是否仍然会放大报告的偏见?来自典型颜色判断的证据
Q3 Environmental Science Pub Date : 2022-09-26 DOI: 10.48550/arXiv.2209.12786
Fangyu Liu, Julian Martin Eisenschlos, Jeremy R. Cole, Nigel Collier
Language models (LMs) trained on raw texts have no direct access to the physical world. Gordon and Van Durme (2013) point out that LMs can thus suffer from reporting bias: texts rarely report on common facts, instead focusing on the unusual aspects of a situation. If LMs are only trained on text corpora and naively memorise local co-occurrence statistics, they thus naturally would learn a biased view of the physical world. While prior studies have repeatedly verified that LMs of smaller scales (e.g., RoBERTa, GPT-2) amplify reporting bias, it remains unknown whether such trends continue when models are scaled up. We investigate reporting bias from the perspective of colour in larger language models (LLMs) such as PaLM and GPT-3. Specifically, we query LLMs for the typical colour of objects, which is one simple type of perceptually grounded physical common sense. Surprisingly, we find that LLMs significantly outperform smaller LMs in determining an object’s typical colour and more closely track human judgments, instead of overfitting to surface patterns stored in texts. This suggests that very large models of language alone are able to overcome certain types of reporting bias that are characterized by local co-occurrences.
在原始文本上训练的语言模型(LMs)无法直接访问物理世界。Gordon和Van Durme(2013)指出,LMs因此可能遭受报道偏见:文本很少报道常见事实,而是关注情况的不寻常方面。如果LMs只在文本语料库上进行训练,并天真地记忆局部共现统计数据,那么它们自然会对物理世界产生偏见。虽然先前的研究已经反复证实,较小规模的LMs(例如RoBERTa, GPT-2)会放大报告偏差,但当模型扩大时,这种趋势是否会继续,仍不得而知。我们从更大的语言模型(llm)如PaLM和GPT-3的颜色角度研究报告偏差。具体来说,我们向llm查询对象的典型颜色,这是一种简单的基于感知的物理常识。令人惊讶的是,我们发现llm在确定物体的典型颜色方面明显优于较小的lm,并且更密切地跟踪人类的判断,而不是过度拟合存储在文本中的表面模式。这表明,非常大的语言模型本身就能够克服某些类型的报告偏差,这些偏差以局部共现为特征。
{"title":"Do ever larger octopi still amplify reporting biases? Evidence from judgments of typical colour","authors":"Fangyu Liu, Julian Martin Eisenschlos, Jeremy R. Cole, Nigel Collier","doi":"10.48550/arXiv.2209.12786","DOIUrl":"https://doi.org/10.48550/arXiv.2209.12786","url":null,"abstract":"Language models (LMs) trained on raw texts have no direct access to the physical world. Gordon and Van Durme (2013) point out that LMs can thus suffer from reporting bias: texts rarely report on common facts, instead focusing on the unusual aspects of a situation. If LMs are only trained on text corpora and naively memorise local co-occurrence statistics, they thus naturally would learn a biased view of the physical world. While prior studies have repeatedly verified that LMs of smaller scales (e.g., RoBERTa, GPT-2) amplify reporting bias, it remains unknown whether such trends continue when models are scaled up. We investigate reporting bias from the perspective of colour in larger language models (LLMs) such as PaLM and GPT-3. Specifically, we query LLMs for the typical colour of objects, which is one simple type of perceptually grounded physical common sense. Surprisingly, we find that LLMs significantly outperform smaller LMs in determining an object’s typical colour and more closely track human judgments, instead of overfitting to surface patterns stored in texts. This suggests that very large models of language alone are able to overcome certain types of reporting bias that are characterized by local co-occurrences.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"18 1","pages":"210-220"},"PeriodicalIF":0.0,"publicationDate":"2022-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80022164","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Re-contextualizing Fairness in NLP: The Case of India 重新界定NLP中的公平性:以印度为例
Q3 Environmental Science Pub Date : 2022-09-25 DOI: 10.48550/arXiv.2209.12226
Shaily Bhatt, Sunipa Dev, Partha P. Talukdar, Shachi Dave, Vinodkumar Prabhakaran
Recent research has revealed undesirable biases in NLP data and models. However, these efforts focus of social disparities in West, and are not directly portable to other geo-cultural contexts. In this paper, we focus on NLP fairness in the context of India. We start with a brief account of the prominent axes of social disparities in India. We build resources for fairness evaluation in the Indian context and use them to demonstrate prediction biases along some of the axes. We then delve deeper into social stereotypes for Region and Religion, demonstrating its prevalence in corpora and models. Finally, we outline a holistic research agenda to re-contextualize NLP fairness research for the Indian context, accounting for Indian societal context, bridging technological gaps in NLP capabilities and resources, and adapting to Indian cultural values. While we focus on India, this framework can be generalized to other geo-cultural contexts.
最近的研究揭示了NLP数据和模型中的不良偏差。然而,这些努力主要集中在西方的社会差异,并不能直接适用于其他地缘文化背景。在本文中,我们关注印度背景下的NLP公平性。我们首先简要介绍一下印度社会差异的主要轴线。我们在印度的背景下建立了公平评估的资源,并用它们来展示沿某些轴的预测偏差。然后,我们深入研究了区域和宗教的社会刻板印象,证明了其在语料库和模型中的普遍存在。最后,我们概述了一个整体的研究议程,以重新定位印度背景下的NLP公平性研究,考虑印度的社会背景,弥合NLP能力和资源方面的技术差距,并适应印度的文化价值观。虽然我们关注的是印度,但这个框架可以推广到其他地缘文化背景。
{"title":"Re-contextualizing Fairness in NLP: The Case of India","authors":"Shaily Bhatt, Sunipa Dev, Partha P. Talukdar, Shachi Dave, Vinodkumar Prabhakaran","doi":"10.48550/arXiv.2209.12226","DOIUrl":"https://doi.org/10.48550/arXiv.2209.12226","url":null,"abstract":"Recent research has revealed undesirable biases in NLP data and models. However, these efforts focus of social disparities in West, and are not directly portable to other geo-cultural contexts. In this paper, we focus on NLP fairness in the context of India. We start with a brief account of the prominent axes of social disparities in India. We build resources for fairness evaluation in the Indian context and use them to demonstrate prediction biases along some of the axes. We then delve deeper into social stereotypes for Region and Religion, demonstrating its prevalence in corpora and models. Finally, we outline a holistic research agenda to re-contextualize NLP fairness research for the Indian context, accounting for Indian societal context, bridging technological gaps in NLP capabilities and resources, and adapting to Indian cultural values. While we focus on India, this framework can be generalized to other geo-cultural contexts.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"514 1","pages":"727-740"},"PeriodicalIF":0.0,"publicationDate":"2022-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77353966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
期刊
AACL Bioflux
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1