首页 > 最新文献

AACL Bioflux最新文献

英文 中文
Dead or Murdered? Predicting Responsibility Perception in Femicide News Reports 死亡还是谋杀?预测杀害女性新闻报道中的责任认知
Q3 Environmental Science Pub Date : 2022-09-24 DOI: 10.48550/arXiv.2209.12030
Gosse Minnema, Sara Gemelli, C. Zanchi, T. Caselli, M. Nissim
Different linguistic expressions can conceptualize the same event from different viewpoints by emphasizing certain participants over others. Here, we investigate a case where this has social consequences: how do linguistic expressions of gender-based violence (GBV) influence who we perceive as responsible? We build on previous psycholinguistic research in this area and conduct a large-scale perception survey of GBV descriptions automatically extracted from a corpus of Italian newspapers. We then train regression models that predict the salience of GBV participants with respect to different dimensions of perceived responsibility. Our best model (fine-tuned BERT) shows solid overall performance, with large differences between dimensions and participants: salient _focus_ is more predictable than salient _blame_, and perpetrators’ salience is more predictable than victims’ salience. Experiments with ridge regression models using different representations show that features based on linguistic theory similarly to word-based features. Overall, we show that different linguistic choices do trigger different perceptions of responsibility, and that such perceptions can be modelled automatically. This work can be a core instrument to raise awareness of the consequences of different perspectivizations in the general public and in news producers alike.
不同的语言表达可以通过强调某些参与者而不是其他参与者,从不同的角度概念化同一事件。在这里,我们调查了一个具有社会后果的案例:基于性别的暴力(GBV)的语言表达如何影响我们认为应该负责的人?我们在这一领域先前的心理语言学研究的基础上,对从意大利报纸语料库中自动提取的GBV描述进行了大规模的感知调查。然后,我们训练回归模型来预测GBV参与者在感知责任的不同维度上的显著性。我们最好的模型(微调BERT)显示出稳定的整体表现,在维度和参与者之间有很大的差异:突出的焦点比突出的责备更容易预测,肇事者的突出比受害者的突出更容易预测。使用不同表示的脊回归模型的实验表明,基于语言学理论的特征与基于单词的特征相似。总的来说,我们表明,不同的语言选择确实会引发不同的责任感知,而且这种感知可以自动建模。这项工作可以成为提高公众和新闻制作人对不同视角后果的认识的核心工具。
{"title":"Dead or Murdered? Predicting Responsibility Perception in Femicide News Reports","authors":"Gosse Minnema, Sara Gemelli, C. Zanchi, T. Caselli, M. Nissim","doi":"10.48550/arXiv.2209.12030","DOIUrl":"https://doi.org/10.48550/arXiv.2209.12030","url":null,"abstract":"Different linguistic expressions can conceptualize the same event from different viewpoints by emphasizing certain participants over others. Here, we investigate a case where this has social consequences: how do linguistic expressions of gender-based violence (GBV) influence who we perceive as responsible? We build on previous psycholinguistic research in this area and conduct a large-scale perception survey of GBV descriptions automatically extracted from a corpus of Italian newspapers. We then train regression models that predict the salience of GBV participants with respect to different dimensions of perceived responsibility. Our best model (fine-tuned BERT) shows solid overall performance, with large differences between dimensions and participants: salient _focus_ is more predictable than salient _blame_, and perpetrators’ salience is more predictable than victims’ salience. Experiments with ridge regression models using different representations show that features based on linguistic theory similarly to word-based features. Overall, we show that different linguistic choices do trigger different perceptions of responsibility, and that such perceptions can be modelled automatically. This work can be a core instrument to raise awareness of the consequences of different perspectivizations in the general public and in news producers alike.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"230 1","pages":"1078-1090"},"PeriodicalIF":0.0,"publicationDate":"2022-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86803013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Whodunit? Learning to Contrast for Authorship Attribution 侦探小说吗?学习对比作者归属
Q3 Environmental Science Pub Date : 2022-09-23 DOI: 10.48550/arXiv.2209.11887
Bo Ai, Yuchen Wang, Yugin Tan, Samson Tan
Authorship attribution is the task of identifying the author of a given text. The key is finding representations that can differentiate between authors. Existing approaches typically use manually designed features that capture a dataset’s content and style, but these approaches are dataset-dependent and yield inconsistent performance across corpora. In this work, we propose to learn author-specific representations by fine-tuning pre-trained generic language representations with a contrastive objective (Contra-X). We show that Contra-X learns representations that form highly separable clusters for different authors. It advances the state-of-the-art on multiple human and machine authorship attribution benchmarks, enabling improvements of up to 6.8% over cross-entropy fine-tuning. However, we find that Contra-X improves overall accuracy at the cost of sacrificing performance for some authors. Resolving this tension will be an important direction for future work. To the best of our knowledge, we are the first to integrate contrastive learning with pre-trained language model fine-tuning for authorship attribution.
作者归属是识别给定文本的作者的任务。关键是找到能够区分作者的表现形式。现有的方法通常使用手动设计的功能来捕获数据集的内容和风格,但是这些方法依赖于数据集,并且在语料库中产生不一致的性能。在这项工作中,我们建议通过微调带有对比目标(contro - x)的预训练通用语言表示来学习作者特定的表示。我们展示了contro - x学习了为不同作者形成高度可分离簇的表示。它在多个人类和机器作者归属基准上推进了最先进的技术,比交叉熵微调提高了6.8%。然而,我们发现contro - x以牺牲某些作者的性能为代价提高了整体准确性。解决这一矛盾将是今后工作的重要方向。据我们所知,我们是第一个将对比学习与预先训练的语言模型微调结合起来的作者归属。
{"title":"Whodunit? Learning to Contrast for Authorship Attribution","authors":"Bo Ai, Yuchen Wang, Yugin Tan, Samson Tan","doi":"10.48550/arXiv.2209.11887","DOIUrl":"https://doi.org/10.48550/arXiv.2209.11887","url":null,"abstract":"Authorship attribution is the task of identifying the author of a given text. The key is finding representations that can differentiate between authors. Existing approaches typically use manually designed features that capture a dataset’s content and style, but these approaches are dataset-dependent and yield inconsistent performance across corpora. In this work, we propose to learn author-specific representations by fine-tuning pre-trained generic language representations with a contrastive objective (Contra-X). We show that Contra-X learns representations that form highly separable clusters for different authors. It advances the state-of-the-art on multiple human and machine authorship attribution benchmarks, enabling improvements of up to 6.8% over cross-entropy fine-tuning. However, we find that Contra-X improves overall accuracy at the cost of sacrificing performance for some authors. Resolving this tension will be an important direction for future work. To the best of our knowledge, we are the first to integrate contrastive learning with pre-trained language model fine-tuning for authorship attribution.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"34 1","pages":"1142-1157"},"PeriodicalIF":0.0,"publicationDate":"2022-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84023264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Learning Interpretable Latent Dialogue Actions With Less Supervision 在较少监督下学习可解释的潜在对话动作
Q3 Environmental Science Pub Date : 2022-09-22 DOI: 10.48550/arXiv.2209.11128
Vojtvech Hudevcek, Ondrej Dusek
We present a novel architecture for explainable modeling of task-oriented dialogues with discrete latent variables to represent dialogue actions. Our model is based on variational recurrent neural networks (VRNN) and requires no explicit annotation of semantic information. Unlike previous works, our approach models the system and user turns separately and performs database query modeling, which makes the model applicable to task-oriented dialogues while producing easily interpretable action latent variables. We show that our model outperforms previous approaches with less supervision in terms of perplexity and BLEU on three datasets, and we propose a way to measure dialogue success without the need for expert annotation. Finally, we propose a novel way to explain semantics of the latent variables with respect to system actions.
我们提出了一种新的架构,用于面向任务的对话的可解释建模,该对话具有离散的潜在变量来表示对话动作。该模型基于变分递归神经网络(VRNN),不需要对语义信息进行显式标注。与以前的工作不同,我们的方法分别为系统和用户转弯建模,并执行数据库查询建模,这使得模型适用于面向任务的对话,同时产生易于解释的动作潜在变量。我们表明,在三个数据集上,我们的模型在困惑度和BLEU方面优于之前较少监督的方法,并且我们提出了一种不需要专家注释来衡量对话成功的方法。最后,我们提出了一种新的方法来解释相对于系统动作的潜在变量的语义。
{"title":"Learning Interpretable Latent Dialogue Actions With Less Supervision","authors":"Vojtvech Hudevcek, Ondrej Dusek","doi":"10.48550/arXiv.2209.11128","DOIUrl":"https://doi.org/10.48550/arXiv.2209.11128","url":null,"abstract":"We present a novel architecture for explainable modeling of task-oriented dialogues with discrete latent variables to represent dialogue actions. Our model is based on variational recurrent neural networks (VRNN) and requires no explicit annotation of semantic information. Unlike previous works, our approach models the system and user turns separately and performs database query modeling, which makes the model applicable to task-oriented dialogues while producing easily interpretable action latent variables. We show that our model outperforms previous approaches with less supervision in terms of perplexity and BLEU on three datasets, and we propose a way to measure dialogue success without the need for expert annotation. Finally, we propose a novel way to explain semantics of the latent variables with respect to system actions.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"16 1","pages":"297-308"},"PeriodicalIF":0.0,"publicationDate":"2022-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85813092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Seeking Diverse Reasoning Logic: Controlled Equation Expression Generation for Solving Math Word Problems 寻求多元推理逻辑:求解数学字题的控制方程表达式生成
Q3 Environmental Science Pub Date : 2022-09-21 DOI: 10.48550/arXiv.2209.10310
Yibin Shen, Qianying Liu, Zhuoyuan Mao, Zhen Wan, Fei Cheng, S. Kurohashi
To solve Math Word Problems, human students leverage diverse reasoning logic that reaches different possible equation solutions. However, the mainstream sequence-to-sequence approach of automatic solvers aims to decode a fixed solution equation supervised by human annotation. In this paper, we propose a controlled equation generation solver by leveraging a set of control codes to guide the model to consider certain reasoning logic and decode the corresponding equations expressions transformed from the human reference. The empirical results suggest that our method universally improves the performance on single-unknown (Math23K) and multiple-unknown (DRAW1K, HMWP) benchmarks, with substantial improvements up to 13.2% accuracy on the challenging multiple-unknown datasets.
为了解决数学应用题,人类学生利用不同的推理逻辑来达到不同的可能的方程解。然而,自动求解器的主流序列到序列方法旨在解码由人类注释监督的固定解方程。在本文中,我们提出了一种控制方程生成求解器,利用一组控制代码来引导模型考虑一定的推理逻辑,并解码由人类参考转换而来的相应方程表达式。实证结果表明,我们的方法普遍提高了单未知(Math23K)和多未知(DRAW1K, HMWP)基准测试的性能,在具有挑战性的多未知数据集上,准确率提高了13.2%。
{"title":"Seeking Diverse Reasoning Logic: Controlled Equation Expression Generation for Solving Math Word Problems","authors":"Yibin Shen, Qianying Liu, Zhuoyuan Mao, Zhen Wan, Fei Cheng, S. Kurohashi","doi":"10.48550/arXiv.2209.10310","DOIUrl":"https://doi.org/10.48550/arXiv.2209.10310","url":null,"abstract":"To solve Math Word Problems, human students leverage diverse reasoning logic that reaches different possible equation solutions. However, the mainstream sequence-to-sequence approach of automatic solvers aims to decode a fixed solution equation supervised by human annotation. In this paper, we propose a controlled equation generation solver by leveraging a set of control codes to guide the model to consider certain reasoning logic and decode the corresponding equations expressions transformed from the human reference. The empirical results suggest that our method universally improves the performance on single-unknown (Math23K) and multiple-unknown (DRAW1K, HMWP) benchmarks, with substantial improvements up to 13.2% accuracy on the challenging multiple-unknown datasets.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"68 1","pages":"254-260"},"PeriodicalIF":0.0,"publicationDate":"2022-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79113952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Dodging the Data Bottleneck: Automatic Subtitling with Automatically Segmented ST Corpora 避开数据瓶颈:基于自动分割语料库的自动字幕
Q3 Environmental Science Pub Date : 2022-09-21 DOI: 10.48550/arXiv.2209.10608
Sara Papi, Alina Karakanta, Matteo Negri, M. Turchi
Speech translation for subtitling (SubST) is the task of automatically translating speech data into well-formed subtitles by inserting subtitle breaks compliant to specific displaying guidelines. Similar to speech translation (ST), model training requires parallel data comprising audio inputs paired with their textual translations. In SubST, however, the text has to be also annotated with subtitle breaks. So far, this requirement has represented a bottleneck for system development, as confirmed by the dearth of publicly available SubST corpora. To fill this gap, we propose a method to convert existing ST corpora into SubST resources without human intervention. We build a segmenter model that automatically segments texts into proper subtitles by exploiting audio and text in a multimodal fashion, achieving high segmentation quality in zero-shot conditions. Comparative experiments with SubST systems respectively trained on manual and automatic segmentations result in similar performance, showing the effectiveness of our approach.
语音字幕翻译(SubST)是通过插入符合特定显示准则的字幕断续符,自动将语音数据翻译成格式良好的字幕的任务。与语音翻译(ST)类似,模型训练需要并行数据,包括与其文本翻译配对的音频输入。然而,在SubST中,文本也必须用副标题分隔符进行注释。到目前为止,这一需求代表了系统开发的瓶颈,因为缺乏公开可用的SubST语料库。为了填补这一空白,我们提出了一种将现有的ST语料库转换为SubST资源而无需人工干预的方法。我们建立了一个切分模型,该模型通过以多模态方式利用音频和文本自动将文本切分为适当的字幕,在零镜头条件下实现高切分质量。与人工和自动分割训练的SubST系统的对比实验结果相似,表明了我们的方法的有效性。
{"title":"Dodging the Data Bottleneck: Automatic Subtitling with Automatically Segmented ST Corpora","authors":"Sara Papi, Alina Karakanta, Matteo Negri, M. Turchi","doi":"10.48550/arXiv.2209.10608","DOIUrl":"https://doi.org/10.48550/arXiv.2209.10608","url":null,"abstract":"Speech translation for subtitling (SubST) is the task of automatically translating speech data into well-formed subtitles by inserting subtitle breaks compliant to specific displaying guidelines. Similar to speech translation (ST), model training requires parallel data comprising audio inputs paired with their textual translations. In SubST, however, the text has to be also annotated with subtitle breaks. So far, this requirement has represented a bottleneck for system development, as confirmed by the dearth of publicly available SubST corpora. To fill this gap, we propose a method to convert existing ST corpora into SubST resources without human intervention. We build a segmenter model that automatically segments texts into proper subtitles by exploiting audio and text in a multimodal fashion, achieving high segmentation quality in zero-shot conditions. Comparative experiments with SubST systems respectively trained on manual and automatic segmentations result in similar performance, showing the effectiveness of our approach.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"108 1","pages":"480-487"},"PeriodicalIF":0.0,"publicationDate":"2022-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86258928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
F-coref: Fast, Accurate and Easy to Use Coreference Resolution F-coref:快速,准确和易于使用的共同参考分辨率
Q3 Environmental Science Pub Date : 2022-09-09 DOI: 10.48550/arXiv.2209.04280
Shon Otmazgin, Arie Cattan, Yoav Goldberg
We introduce fastcoref, a python package for fast, accurate, and easy-to-use English coreference resolution. The package is pip-installable, and allows two modes: an accurate mode based on the LingMess architecture, providing state-of-the-art coreference accuracy, and a substantially faster model, F-coref, which is the focus of this work. F-coref allows to process 2.8K OntoNotes documents in 25 seconds on a V100 GPU (compared to 6 minutes for the LingMess model, and to 12 minutes of the popular AllenNLP coreference model) with only a modest drop in accuracy. The fast speed is achieved through a combination of distillation of a compact model from the LingMess model, and an efficient batching implementation using a technique we call leftover batching. https://github.com/shon-otmazgin/fastcoref
我们介绍fastcoref,这是一个python包,用于快速、准确和易于使用的英语共同参考解析。该包是可pip安装的,并允许两种模式:基于LingMess架构的精确模式,提供最先进的共参考精度,以及更快的模型F-coref,这是本工作的重点。F-coref允许在V100 GPU上在25秒内处理2.8K OntoNotes文档(相比之下,LingMess模型需要6分钟,流行的AllenNLP共同参考模型需要12分钟),精度只有轻微下降。快速的速度是通过结合LingMess模型的精简模型的蒸馏,以及使用我们称之为剩余批处理技术的高效批处理实现来实现的。https://github.com/shon-otmazgin/fastcoref
{"title":"F-coref: Fast, Accurate and Easy to Use Coreference Resolution","authors":"Shon Otmazgin, Arie Cattan, Yoav Goldberg","doi":"10.48550/arXiv.2209.04280","DOIUrl":"https://doi.org/10.48550/arXiv.2209.04280","url":null,"abstract":"We introduce fastcoref, a python package for fast, accurate, and easy-to-use English coreference resolution. The package is pip-installable, and allows two modes: an accurate mode based on the LingMess architecture, providing state-of-the-art coreference accuracy, and a substantially faster model, F-coref, which is the focus of this work. F-coref allows to process 2.8K OntoNotes documents in 25 seconds on a V100 GPU (compared to 6 minutes for the LingMess model, and to 12 minutes of the popular AllenNLP coreference model) with only a modest drop in accuracy. The fast speed is achieved through a combination of distillation of a compact model from the LingMess model, and an efficient batching implementation using a technique we call leftover batching. https://github.com/shon-otmazgin/fastcoref","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"47 1","pages":"48-56"},"PeriodicalIF":0.0,"publicationDate":"2022-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74998841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
UKP-SQuARE v2: Explainability and Adversarial Attacks for Trustworthy QA UKP-SQuARE v2:可解释性和可信赖QA的对抗性攻击
Q3 Environmental Science Pub Date : 2022-08-19 DOI: 10.48550/arXiv.2208.09316
Rachneet Sachdeva, Haritz Puerto, Tim Baumgärtner, Sewin Tariverdian, Hao Zhang, Kexin Wang, H. Saad, Leonardo F. R. Ribeiro, Iryna Gurevych
Question Answering (QA) systems are increasingly deployed in applications where they support real-world decisions. However, state-of-the-art models rely on deep neural networks, which are difficult to interpret by humans. Inherently interpretable models or post hoc explainability methods can help users to comprehend how a model arrives at its prediction and, if successful, increase their trust in the system. Furthermore, researchers can leverage these insights to develop new methods that are more accurate and less biased. In this paper, we introduce SQuARE v2, the new version of SQuARE, to provide an explainability infrastructure for comparing models based on methods such as saliency maps and graph-based explanations. While saliency maps are useful to inspect the importance of each input token for the model’s prediction, graph-based explanations from external Knowledge Graphs enable the users to verify the reasoning behind the model prediction. In addition, we provide multiple adversarial attacks to compare the robustness of QA models. With these explainability methods and adversarial attacks, we aim to ease the research on trustworthy QA models. SQuARE is available on https://square.ukp-lab.de.
问答系统越来越多地部署在支持实际决策的应用程序中。然而,最先进的模型依赖于人类难以理解的深度神经网络。固有的可解释模型或事后可解释性方法可以帮助用户理解模型是如何实现其预测的,如果成功,可以增加他们对系统的信任。此外,研究人员可以利用这些见解来开发更准确、更少偏见的新方法。在本文中,我们引入了SQuARE的新版本SQuARE v2,为基于显著性图和基于图的解释等方法的模型比较提供了一个可解释性基础设施。虽然显著性图对于检查模型预测的每个输入标记的重要性很有用,但来自外部知识图的基于图的解释使用户能够验证模型预测背后的推理。此外,我们提供了多个对抗性攻击来比较QA模型的鲁棒性。利用这些可解释性方法和对抗性攻击,我们的目标是简化可信质量保证模型的研究。SQuARE可以在https://square.ukp-lab.de上找到。
{"title":"UKP-SQuARE v2: Explainability and Adversarial Attacks for Trustworthy QA","authors":"Rachneet Sachdeva, Haritz Puerto, Tim Baumgärtner, Sewin Tariverdian, Hao Zhang, Kexin Wang, H. Saad, Leonardo F. R. Ribeiro, Iryna Gurevych","doi":"10.48550/arXiv.2208.09316","DOIUrl":"https://doi.org/10.48550/arXiv.2208.09316","url":null,"abstract":"Question Answering (QA) systems are increasingly deployed in applications where they support real-world decisions. However, state-of-the-art models rely on deep neural networks, which are difficult to interpret by humans. Inherently interpretable models or post hoc explainability methods can help users to comprehend how a model arrives at its prediction and, if successful, increase their trust in the system. Furthermore, researchers can leverage these insights to develop new methods that are more accurate and less biased. In this paper, we introduce SQuARE v2, the new version of SQuARE, to provide an explainability infrastructure for comparing models based on methods such as saliency maps and graph-based explanations. While saliency maps are useful to inspect the importance of each input token for the model’s prediction, graph-based explanations from external Knowledge Graphs enable the users to verify the reasoning behind the model prediction. In addition, we provide multiple adversarial attacks to compare the robustness of QA models. With these explainability methods and adversarial attacks, we aim to ease the research on trustworthy QA models. SQuARE is available on https://square.ukp-lab.de.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"2 1","pages":"28-38"},"PeriodicalIF":0.0,"publicationDate":"2022-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88122488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
TexPrax: A Messaging Application for Ethical, Real-time Data Collection and Annotation TexPrax:一个道德的、实时的数据收集和注释的消息传递应用程序
Q3 Environmental Science Pub Date : 2022-08-16 DOI: 10.48550/arXiv.2208.07846
Lorenz Stangier, Ji-Ung Lee, Yuxi Wang, Marvin Müller, Nicholas R. J. Frick, J. Metternich, Iryna Gurevych
Collecting and annotating task-oriented dialog data is difficult, especially for highly specific domains that require expert knowledge. At the same time, informal communication channels such as instant messengers are increasingly being used at work. This has led to a lot of work-relevant information that is disseminated through those channels and needs to be post-processed manually by the employees. To alleviate this problem, we present TexPrax, a messaging system to collect and annotate _problems_, _causes_, and _solutions_ that occur in work-related chats. TexPrax uses a chatbot to directly engage the employees to provide lightweight annotations on their conversation and ease their documentation work. To comply with data privacy and security regulations, we use an end-to-end message encryption and give our users full control over their data which has various advantages over conventional annotation tools. We evaluate TexPrax in a user-study with German factory employees who ask their colleagues for solutions on problems that arise during their daily work. Overall, we collect 202 task-oriented German dialogues containing 1,027 sentences with sentence-level expert annotations. Our data analysis also reveals that real-world conversations frequently contain instances with code-switching, varying abbreviations for the same entity, and dialects which NLP systems should be able to handle.
收集和注释面向任务的对话框数据是困难的,特别是对于需要专家知识的高度特定的领域。与此同时,诸如即时通讯工具之类的非正式沟通渠道也越来越多地用于工作中。这导致大量与工作相关的信息通过这些渠道传播,需要员工手工进行后期处理。为了缓解这个问题,我们介绍了TexPrax,这是一个消息传递系统,用于收集和注释与工作相关的聊天中出现的问题、原因和解决方案。TexPrax使用聊天机器人直接吸引员工,为他们的对话提供轻量级注释,并简化他们的文档工作。为了遵守数据隐私和安全法规,我们使用端到端消息加密,让我们的用户完全控制他们的数据,这比传统的注释工具有很多优点。我们在对德国工厂员工的用户研究中对TexPrax进行了评估,这些员工在日常工作中遇到问题时会向同事寻求解决方案。总的来说,我们收集了202个面向任务的德语对话,其中包含1,027个句子,其中包含句子级专家注释。我们的数据分析还表明,现实世界的对话经常包含代码切换的实例,同一实体的不同缩写,以及NLP系统应该能够处理的方言。
{"title":"TexPrax: A Messaging Application for Ethical, Real-time Data Collection and Annotation","authors":"Lorenz Stangier, Ji-Ung Lee, Yuxi Wang, Marvin Müller, Nicholas R. J. Frick, J. Metternich, Iryna Gurevych","doi":"10.48550/arXiv.2208.07846","DOIUrl":"https://doi.org/10.48550/arXiv.2208.07846","url":null,"abstract":"Collecting and annotating task-oriented dialog data is difficult, especially for highly specific domains that require expert knowledge. At the same time, informal communication channels such as instant messengers are increasingly being used at work. This has led to a lot of work-relevant information that is disseminated through those channels and needs to be post-processed manually by the employees. To alleviate this problem, we present TexPrax, a messaging system to collect and annotate _problems_, _causes_, and _solutions_ that occur in work-related chats. TexPrax uses a chatbot to directly engage the employees to provide lightweight annotations on their conversation and ease their documentation work. To comply with data privacy and security regulations, we use an end-to-end message encryption and give our users full control over their data which has various advantages over conventional annotation tools. We evaluate TexPrax in a user-study with German factory employees who ask their colleagues for solutions on problems that arise during their daily work. Overall, we collect 202 task-oriented German dialogues containing 1,027 sentences with sentence-level expert annotations. Our data analysis also reveals that real-world conversations frequently contain instances with code-switching, varying abbreviations for the same entity, and dialects which NLP systems should be able to handle.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"6 1","pages":"9-16"},"PeriodicalIF":0.0,"publicationDate":"2022-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80374544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
PInKS: Preconditioned Commonsense Inference with Minimal Supervision 粉红色:具有最小监督的预设常识推理
Q3 Environmental Science Pub Date : 2022-06-16 DOI: 10.48550/arXiv.2206.07920
Ehsan Qasemi, Piyush Khanna, Qiang Ning, Muhao Chen
Reasoning with preconditions such as “glass can be used for drinking water unless the glass is shattered” remains an open problem for language models. The main challenge lies in the scarcity of preconditions data and the model’s lack of support for such reasoning. We present PInKS , Preconditioned Commonsense Inference with WeaK Supervision, an improved model for reasoning with preconditions through minimum supervision. We show, empirically and theoretically, that PInKS improves the results on benchmarks focused on reasoning with the preconditions of commonsense knowledge (up to 40% Macro-F1 scores). We further investigate PInKS through PAC-Bayesian informativeness analysis, precision measures, and ablation study.
对语言模型来说,用诸如“玻璃可以用来喝水,除非玻璃碎了”这样的前提条件进行推理仍然是一个有待解决的问题。主要的挑战在于先决条件数据的稀缺性和模型缺乏对这种推理的支持。我们提出了PInKS, Preconditioned Commonsense Inference with WeaK Supervision,这是一个通过最小监督进行有前提条件推理的改进模型。我们从经验和理论上证明,PInKS提高了以常识知识为前提的推理基准测试的结果(高达40%的Macro-F1分数)。我们通过PAC-Bayesian信息性分析、精度测量和消融研究进一步研究了PInKS。
{"title":"PInKS: Preconditioned Commonsense Inference with Minimal Supervision","authors":"Ehsan Qasemi, Piyush Khanna, Qiang Ning, Muhao Chen","doi":"10.48550/arXiv.2206.07920","DOIUrl":"https://doi.org/10.48550/arXiv.2206.07920","url":null,"abstract":"Reasoning with preconditions such as “glass can be used for drinking water unless the glass is shattered” remains an open problem for language models. The main challenge lies in the scarcity of preconditions data and the model’s lack of support for such reasoning. We present PInKS , Preconditioned Commonsense Inference with WeaK Supervision, an improved model for reasoning with preconditions through minimum supervision. We show, empirically and theoretically, that PInKS improves the results on benchmarks focused on reasoning with the preconditions of commonsense knowledge (up to 40% Macro-F1 scores). We further investigate PInKS through PAC-Bayesian informativeness analysis, precision measures, and ablation study.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"6 1","pages":"320-336"},"PeriodicalIF":0.0,"publicationDate":"2022-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87035369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Director: Generator-Classifiers For Supervised Language Modeling 主管:监督语言建模的生成器-分类器
Q3 Environmental Science Pub Date : 2022-06-15 DOI: 10.48550/arXiv.2206.07694
Kushal Arora, Kurt Shuster, Sainbayar Sukhbaatar, J. Weston
Current language models achieve low perplexity but their resulting generations still suffer from toxic responses, repetitiveness, and contradictions. The standard language modeling setup fails to address these issues. In this paper, we introduce a new architecture, Director, that consists of a unified generator-classifier with both a language modeling and a classification head for each output token. Training is conducted jointly using both standard language modeling data, and data labeled with desirable and undesirable sequences. Experiments in several settings show that the model has competitive training and decoding speed compared to standard language models while yielding superior results, avoiding undesirable behaviors while maintaining generation quality. It also outperforms existing model guiding approaches in terms of both accuracy and efficiency. Our code is made publicly available.
目前的语言模型达到了较低的困惑度,但它们产生的后代仍然受到有毒反应、重复和矛盾的影响。标准的语言建模设置无法解决这些问题。在本文中,我们介绍了一个新的体系结构Director,它由一个统一的生成器-分类器组成,每个输出标记都有一个语言建模和分类头。训练是联合使用标准语言建模数据和标记了理想和不理想序列的数据进行的。在几种环境下的实验表明,与标准语言模型相比,该模型具有竞争性的训练和解码速度,同时产生更好的结果,避免了不良行为,同时保持了生成质量。它在准确性和效率方面也优于现有的模型指导方法。我们的代码是公开的。
{"title":"Director: Generator-Classifiers For Supervised Language Modeling","authors":"Kushal Arora, Kurt Shuster, Sainbayar Sukhbaatar, J. Weston","doi":"10.48550/arXiv.2206.07694","DOIUrl":"https://doi.org/10.48550/arXiv.2206.07694","url":null,"abstract":"Current language models achieve low perplexity but their resulting generations still suffer from toxic responses, repetitiveness, and contradictions. The standard language modeling setup fails to address these issues. In this paper, we introduce a new architecture, Director, that consists of a unified generator-classifier with both a language modeling and a classification head for each output token. Training is conducted jointly using both standard language modeling data, and data labeled with desirable and undesirable sequences. Experiments in several settings show that the model has competitive training and decoding speed compared to standard language models while yielding superior results, avoiding undesirable behaviors while maintaining generation quality. It also outperforms existing model guiding approaches in terms of both accuracy and efficiency. Our code is made publicly available.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"26 1","pages":"512-526"},"PeriodicalIF":0.0,"publicationDate":"2022-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83068655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
期刊
AACL Bioflux
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1