首页 > 最新文献

AACL Bioflux最新文献

英文 中文
Phylogeny-Inspired Adaptation of Multilingual Models to New Languages 多语言模式对新语言的系统发育启发适应
Q3 Environmental Science Pub Date : 2022-05-19 DOI: 10.48550/arXiv.2205.09634
FAHIM FAISAL, Antonios Anastasopoulos
Large pretrained multilingual models, trained on dozens of languages, have delivered promising results due to cross-lingual learning capabilities on a variety of language tasks. Further adapting these models to specific languages, especially ones unseen during pre-training, is an important goal toward expanding the coverage of language technologies. In this study, we show how we can use language phylogenetic information to improve cross-lingual transfer leveraging closely related languages in a structured, linguistically-informed manner. We perform adapter-based training on languages from diverse language families (Germanic, Uralic, Tupian, Uto-Aztecan) and evaluate on both syntactic and semantic tasks, obtaining more than 20% relative performance improvements over strong commonly used baselines, especially on languages unseen during pre-training.
大型预训练的多语言模型,经过数十种语言的训练,由于在各种语言任务上的跨语言学习能力,已经取得了很好的结果。进一步使这些模型适应特定的语言,特别是那些在预训练期间未见过的语言,是扩大语言技术覆盖范围的一个重要目标。在这项研究中,我们展示了如何利用语言系统发育信息,以结构化的、语言知情的方式利用密切相关的语言来改善跨语言迁移。我们对来自不同语系的语言(日耳曼语、乌拉尔语、图pian、乌托-阿兹特克语)进行了基于适配器的训练,并对句法和语义任务进行了评估,在强大的常用基线上获得了超过20%的相对性能提升,特别是在预训练期间未见过的语言上。
{"title":"Phylogeny-Inspired Adaptation of Multilingual Models to New Languages","authors":"FAHIM FAISAL, Antonios Anastasopoulos","doi":"10.48550/arXiv.2205.09634","DOIUrl":"https://doi.org/10.48550/arXiv.2205.09634","url":null,"abstract":"Large pretrained multilingual models, trained on dozens of languages, have delivered promising results due to cross-lingual learning capabilities on a variety of language tasks. Further adapting these models to specific languages, especially ones unseen during pre-training, is an important goal toward expanding the coverage of language technologies. In this study, we show how we can use language phylogenetic information to improve cross-lingual transfer leveraging closely related languages in a structured, linguistically-informed manner. We perform adapter-based training on languages from diverse language families (Germanic, Uralic, Tupian, Uto-Aztecan) and evaluate on both syntactic and semantic tasks, obtaining more than 20% relative performance improvements over strong commonly used baselines, especially on languages unseen during pre-training.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"74 1","pages":"434-452"},"PeriodicalIF":0.0,"publicationDate":"2022-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81575784","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
A Unified Model for Reverse Dictionary and Definition Modelling 一种统一的反向字典和定义建模模型
Q3 Environmental Science Pub Date : 2022-05-09 DOI: 10.48550/arXiv.2205.04602
Pinzhen Chen, Zheng Zhao
We build a dual-way neural dictionary to retrieve words given definitions, and produce definitions for queried words. The model learns the two tasks simultaneously and handles unknown words via embeddings. It casts a word or a definition to the same representation space through a shared layer, then generates the other form in a multi-task fashion. Our method achieves promising automatic scores on previous benchmarks without extra resources. Human annotators prefer the model’s outputs in both reference-less and reference-based evaluation, indicating its practicality. Analysis suggests that multiple objectives benefit learning.
我们建立了一个双向神经词典来检索给定定义的单词,并为查询的单词生成定义。该模型同时学习两个任务,并通过嵌入处理未知单词。它通过共享层将单词或定义强制转换到相同的表示空间,然后以多任务方式生成另一种形式。我们的方法在没有额外资源的情况下在以前的基准测试中实现了有希望的自动评分。人类注释者在无参考和基于参考的评估中都更喜欢模型的输出,这表明它的实用性。分析表明,多重目标有利于学习。
{"title":"A Unified Model for Reverse Dictionary and Definition Modelling","authors":"Pinzhen Chen, Zheng Zhao","doi":"10.48550/arXiv.2205.04602","DOIUrl":"https://doi.org/10.48550/arXiv.2205.04602","url":null,"abstract":"We build a dual-way neural dictionary to retrieve words given definitions, and produce definitions for queried words. The model learns the two tasks simultaneously and handles unknown words via embeddings. It casts a word or a definition to the same representation space through a shared layer, then generates the other form in a multi-task fashion. Our method achieves promising automatic scores on previous benchmarks without extra resources. Human annotators prefer the model’s outputs in both reference-less and reference-based evaluation, indicating its practicality. Analysis suggests that multiple objectives benefit learning.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"82 1","pages":"8-13"},"PeriodicalIF":0.0,"publicationDate":"2022-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84004930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Prompt Array Keeps the Bias Away: Debiasing Vision-Language Models with Adversarial Learning 提示数组使偏见远离:对抗学习消除视觉语言模型的偏见
Q3 Environmental Science Pub Date : 2022-03-22 DOI: 10.48550/arXiv.2203.11933
Hugo Elias Berg, S. Hall, Yash Bhalgat, Wonsuk Yang, Hannah Rose Kirk, Aleksandar Shtedritski, Max Bain
Vision-language models can encode societal biases and stereotypes, but there are challenges to measuring and mitigating these multimodal harms due to lacking measurement robustness and feature degradation. To address these challenges, we investigate bias measures and apply ranking metrics for image-text representations. We then investigate debiasing methods and show that prepending learned embeddings to text queries that are jointly trained with adversarial debiasing and a contrastive loss, reduces various bias measures with minimal degradation to the image-text representation.
视觉语言模型可以编码社会偏见和刻板印象,但由于缺乏测量鲁棒性和特征退化,在测量和减轻这些多模态危害方面存在挑战。为了解决这些挑战,我们研究了偏见措施,并应用图像-文本表示的排名指标。然后,我们研究了去偏方法,并表明将学习到的嵌入添加到使用对抗性去偏和对比损失联合训练的文本查询中,减少了各种偏差度量,同时最小化了对图像-文本表示的退化。
{"title":"A Prompt Array Keeps the Bias Away: Debiasing Vision-Language Models with Adversarial Learning","authors":"Hugo Elias Berg, S. Hall, Yash Bhalgat, Wonsuk Yang, Hannah Rose Kirk, Aleksandar Shtedritski, Max Bain","doi":"10.48550/arXiv.2203.11933","DOIUrl":"https://doi.org/10.48550/arXiv.2203.11933","url":null,"abstract":"Vision-language models can encode societal biases and stereotypes, but there are challenges to measuring and mitigating these multimodal harms due to lacking measurement robustness and feature degradation. To address these challenges, we investigate bias measures and apply ranking metrics for image-text representations. We then investigate debiasing methods and show that prepending learned embeddings to text queries that are jointly trained with adversarial debiasing and a contrastive loss, reduces various bias measures with minimal degradation to the image-text representation.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"48 1","pages":"806-822"},"PeriodicalIF":0.0,"publicationDate":"2022-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79165225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 40
Seamlessly Integrating Factual Information and Social Content with Persuasive Dialogue 无缝整合事实信息和社会内容与有说服力的对话
Q3 Environmental Science Pub Date : 2022-03-15 DOI: 10.48550/arXiv.2203.07657
Maximillian Chen, Weiyan Shi, Feifan Yan, Ryan Hou, Jingwen Zhang, Saurav Sahay, Zhou Yu
Complex conversation settings such as persuasion involve communicating changes in attitude or behavior, so users’ perspectives need to be addressed, even when not directly related to the topic. In this work, we contribute a novel modular dialogue system framework that seamlessly integrates factual information and social content into persuasive dialogue. Our framework is generalizable to any dialogue tasks that have mixed social and task contents. We conducted a study that compared user evaluations of our framework versus a baseline end-to-end generation model. We found our model was evaluated to be more favorable in all dimensions including competence and friendliness compared to the baseline model which does not explicitly handle social content or factual questions.
复杂的对话设置,如说服,涉及沟通态度或行为的变化,因此需要解决用户的观点,即使与主题没有直接关系。在这项工作中,我们贡献了一个新颖的模块化对话系统框架,将事实信息和社会内容无缝集成到有说服力的对话中。我们的框架可以推广到任何混合了社交和任务内容的对话任务。我们进行了一项研究,将用户对我们框架的评估与基线端到端生成模型进行了比较。我们发现,与没有明确处理社会内容或事实问题的基线模型相比,我们的模型在包括能力和友好性在内的所有维度上都被评估得更有利。
{"title":"Seamlessly Integrating Factual Information and Social Content with Persuasive Dialogue","authors":"Maximillian Chen, Weiyan Shi, Feifan Yan, Ryan Hou, Jingwen Zhang, Saurav Sahay, Zhou Yu","doi":"10.48550/arXiv.2203.07657","DOIUrl":"https://doi.org/10.48550/arXiv.2203.07657","url":null,"abstract":"Complex conversation settings such as persuasion involve communicating changes in attitude or behavior, so users’ perspectives need to be addressed, even when not directly related to the topic. In this work, we contribute a novel modular dialogue system framework that seamlessly integrates factual information and social content into persuasive dialogue. Our framework is generalizable to any dialogue tasks that have mixed social and task contents. We conducted a study that compared user evaluations of our framework versus a baseline end-to-end generation model. We found our model was evaluated to be more favorable in all dimensions including competence and friendliness compared to the baseline model which does not explicitly handle social content or factual questions.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"22 1","pages":"399-413"},"PeriodicalIF":0.0,"publicationDate":"2022-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81482344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A Simple and Effective Usage of Word Clusters for CBOW Model CBOW模型中一种简单有效的词簇用法
Q3 Environmental Science Pub Date : 2020-01-01 DOI: 10.5715/jnlp.29.785
Yukun Feng, Chenlong Hu, Hidetaka Kamigaito, Hiroya Takamura, M. Okumura
We propose a simple and effective method for incorporating word clusters into the Continuous Bag-of-Words (CBOW) model. Specifically, we propose to replace infrequent input and output words in CBOW model with their clusters. The resulting cluster-incorporated CBOW model produces embeddings of frequent words and a small amount of cluster embeddings, which will be fine-tuned in downstream tasks. We empirically show our replacing method works well on several downstream tasks. Through our analysis, we show that our method might be also useful for other similar models which produce word embeddings.
我们提出了一种简单有效的方法将聚类纳入连续词袋模型。具体来说,我们建议将CBOW模型中不频繁的输入输出词替换为它们的聚类。所得到的聚类结合的CBOW模型产生频繁词的嵌入和少量的聚类嵌入,这些嵌入将在后续任务中进行微调。我们的经验表明,我们的替换方法在几个下游任务上工作得很好。通过我们的分析,我们表明我们的方法可能对其他产生词嵌入的类似模型也有用。
{"title":"A Simple and Effective Usage of Word Clusters for CBOW Model","authors":"Yukun Feng, Chenlong Hu, Hidetaka Kamigaito, Hiroya Takamura, M. Okumura","doi":"10.5715/jnlp.29.785","DOIUrl":"https://doi.org/10.5715/jnlp.29.785","url":null,"abstract":"We propose a simple and effective method for incorporating word clusters into the Continuous Bag-of-Words (CBOW) model. Specifically, we propose to replace infrequent input and output words in CBOW model with their clusters. The resulting cluster-incorporated CBOW model produces embeddings of frequent words and a small amount of cluster embeddings, which will be fine-tuned in downstream tasks. We empirically show our replacing method works well on several downstream tasks. Through our analysis, we show that our method might be also useful for other similar models which produce word embeddings.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"42 1","pages":"80-86"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75510058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
You May Like This Hotel Because ...: Identifying Evidence for Explainable Recommendations 你喜欢这家酒店可能是因为……:确定可解释建议的证据
Q3 Environmental Science Pub Date : 2020-01-01 DOI: 10.5715/JNLP.28.264
Shin Kanouchi, Masato Neishi, Yuta Hayashibe, Hiroki Ouchi, Naoaki Okazaki
Explainable recommendation is a good way to improve user satisfaction. However, explainable recommendation in dialogue is challenging since it has to handle natural language as both input and output. To tackle the challenge, this paper proposes a novel and practical task to explain evidences in recommending hotels given vague requests expressed freely in natural language. We decompose the process into two subtasks on hotel reviews: Evidence Identification and Evidence Explanation. The former predicts whether or not a sentence contains evidence that expresses why a given request is satisfied. The latter generates a recommendation sentence given a request and an evidence sentence. In order to address these subtasks, we build an Evidence-based Explanation dataset, which is the largest dataset for explaining evidences in recommending hotels for vague requests. The experimental results demonstrate that the BERT model can find evidence sentences with respect to various vague requests and that the LSTM-based model can generate recommendation sentences.
可解释的推荐是提高用户满意度的好方法。然而,对话中的可解释推荐具有挑战性,因为它必须处理自然语言作为输入和输出。为了应对这一挑战,本文提出了一个新颖而实用的任务,解释在自然语言中自由表达的模糊要求下推荐酒店的证据。我们将酒店评论的过程分解为两个子任务:证据识别和证据解释。前者预测一个句子是否包含表达为什么满足给定请求的证据。后者根据请求生成建议句和证据句。为了解决这些子任务,我们建立了一个基于证据的解释数据集,这是最大的数据集,用于解释为模糊请求推荐酒店的证据。实验结果表明,BERT模型可以针对各种模糊请求找到证据句,基于lstm的模型可以生成推荐句。
{"title":"You May Like This Hotel Because ...: Identifying Evidence for Explainable Recommendations","authors":"Shin Kanouchi, Masato Neishi, Yuta Hayashibe, Hiroki Ouchi, Naoaki Okazaki","doi":"10.5715/JNLP.28.264","DOIUrl":"https://doi.org/10.5715/JNLP.28.264","url":null,"abstract":"Explainable recommendation is a good way to improve user satisfaction. However, explainable recommendation in dialogue is challenging since it has to handle natural language as both input and output. To tackle the challenge, this paper proposes a novel and practical task to explain evidences in recommending hotels given vague requests expressed freely in natural language. We decompose the process into two subtasks on hotel reviews: Evidence Identification and Evidence Explanation. The former predicts whether or not a sentence contains evidence that expresses why a given request is satisfied. The latter generates a recommendation sentence given a request and an evidence sentence. In order to address these subtasks, we build an Evidence-based Explanation dataset, which is the largest dataset for explaining evidences in recommending hotels for vague requests. The experimental results demonstrate that the BERT model can find evidence sentences with respect to various vague requests and that the LSTM-based model can generate recommendation sentences.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"11 1","pages":"890-899"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78349942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
AACL Bioflux
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1