首页 > 最新文献

Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)最新文献

英文 中文
Veritas Annotator: Discovering the Origin of a Rumour 真理注释者:发现谣言的起源
Lucas Azevedo, Mohamed Moustafa
Defined as the intentional or unintentionalspread of false information (K et al., 2019)through context and/or content manipulation,fake news has become one of the most seriousproblems associated with online information(Waldrop, 2017). Consequently, it comes asno surprise that Fake News Detection hasbecome one of the major foci of variousfields of machine learning and while machinelearning models have allowed individualsand companies to automate decision-basedprocesses that were once thought to be onlydoable by humans, it is no secret that thereal-life applications of such models are notviable without the existence of an adequatetraining dataset. In this paper we describethe Veritas Annotator, a web application formanually identifying the origin of a rumour.These rumours, often referred as claims,were previously checked for validity byFact-Checking Agencies.
假新闻被定义为通过语境和/或内容操纵有意或无意地传播虚假信息(K等人,2019),已成为与在线信息相关的最严重问题之一(Waldrop, 2017)。因此,假新闻检测成为机器学习各个领域的主要焦点之一也就不足为奇了,虽然机器学习模型允许个人和公司自动化曾经被认为只能由人类完成的基于决策的过程,但如果没有足够的训练数据集,这些模型的现实应用是不可行的,这已经不是什么秘密了。在本文中,我们描述了Veritas Annotator,一个用于手动识别谣言来源的web应用程序。这些谣言,通常被称为索赔,之前由事实核查机构进行了有效性核查。
{"title":"Veritas Annotator: Discovering the Origin of a Rumour","authors":"Lucas Azevedo, Mohamed Moustafa","doi":"10.18653/v1/D19-6614","DOIUrl":"https://doi.org/10.18653/v1/D19-6614","url":null,"abstract":"Defined as the intentional or unintentionalspread of false information (K et al., 2019)through context and/or content manipulation,fake news has become one of the most seriousproblems associated with online information(Waldrop, 2017). Consequently, it comes asno surprise that Fake News Detection hasbecome one of the major foci of variousfields of machine learning and while machinelearning models have allowed individualsand companies to automate decision-basedprocesses that were once thought to be onlydoable by humans, it is no secret that thereal-life applications of such models are notviable without the existence of an adequatetraining dataset. In this paper we describethe Veritas Annotator, a web application formanually identifying the origin of a rumour.These rumours, often referred as claims,were previously checked for validity byFact-Checking Agencies.","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133517095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Relation Extraction among Multiple Entities Using a Dual Pointer Network with a Multi-Head Attention Mechanism 基于多头注意机制的双指针网络多实体关系提取
Seongsik Park, H. Kim
Many previous studies on relation extrac-tion have been focused on finding only one relation between two entities in a single sentence. However, we can easily find the fact that multiple entities exist in a single sentence and the entities form multiple relations. To resolve this prob-lem, we propose a relation extraction model based on a dual pointer network with a multi-head attention mechanism. The proposed model finds n-to-1 subject-object relations by using a forward de-coder called an object decoder. Then, it finds 1-to-n subject-object relations by using a backward decoder called a sub-ject decoder. In the experiments with the ACE-05 dataset and the NYT dataset, the proposed model achieved the state-of-the-art performances (F1-score of 80.5% in the ACE-05 dataset, F1-score of 78.3% in the NYT dataset)
以往的许多关系提取研究都集中在寻找单个句子中两个实体之间的一个关系。然而,我们很容易发现一个句子中存在多个实体,并且这些实体形成了多种关系。为了解决这一问题,我们提出了一种基于多头注意机制的双指针网络的关系提取模型。提出的模型通过使用称为对象解码器的前向解码器找到n- 1的主题-对象关系。然后,它通过使用称为主题解码器的向后解码器找到1对n的主题-对象关系。在ACE-05数据集和NYT数据集的实验中,所提出的模型达到了最先进的性能(ACE-05数据集的f1得分为80.5%,NYT数据集的f1得分为78.3%)。
{"title":"Relation Extraction among Multiple Entities Using a Dual Pointer Network with a Multi-Head Attention Mechanism","authors":"Seongsik Park, H. Kim","doi":"10.18653/v1/D19-6608","DOIUrl":"https://doi.org/10.18653/v1/D19-6608","url":null,"abstract":"Many previous studies on relation extrac-tion have been focused on finding only one relation between two entities in a single sentence. However, we can easily find the fact that multiple entities exist in a single sentence and the entities form multiple relations. To resolve this prob-lem, we propose a relation extraction model based on a dual pointer network with a multi-head attention mechanism. The proposed model finds n-to-1 subject-object relations by using a forward de-coder called an object decoder. Then, it finds 1-to-n subject-object relations by using a backward decoder called a sub-ject decoder. In the experiments with the ACE-05 dataset and the NYT dataset, the proposed model achieved the state-of-the-art performances (F1-score of 80.5% in the ACE-05 dataset, F1-score of 78.3% in the NYT dataset)","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121010652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Team GPLSI. Approach for automated fact checking 团队GPLSI。自动事实检查的方法
Aimée Alonso-Reina, Robiert Sepúlveda-Torres, E. Saquete, M. Palomar
Fever Shared 2.0 Task is a challenge meant for developing automated fact checking systems. Our approach for the Fever 2.0 is based on a previous proposal developed by Team Athene UKP TU Darmstadt. Our proposal modifies the sentence retrieval phase, using statement extraction and representation in the form of triplets (subject, object, action). Triplets are extracted from the claim and compare to triplets extracted from Wikipedia articles using semantic similarity. Our results are satisfactory but there is room for improvement.
Fever Shared 2.0 Task是一项旨在开发自动事实检查系统的挑战。我们对Fever 2.0的方法是基于先前由Team Athene UKP TU Darmstadt开发的提案。我们的建议修改了句子检索阶段,使用三元组(主语、宾语、动作)形式的语句提取和表示。从声明中提取三元组,并使用语义相似性将其与从维基百科文章中提取的三元组进行比较。我们的结果令人满意,但仍有改进的余地。
{"title":"Team GPLSI. Approach for automated fact checking","authors":"Aimée Alonso-Reina, Robiert Sepúlveda-Torres, E. Saquete, M. Palomar","doi":"10.18653/v1/D19-6617","DOIUrl":"https://doi.org/10.18653/v1/D19-6617","url":null,"abstract":"Fever Shared 2.0 Task is a challenge meant for developing automated fact checking systems. Our approach for the Fever 2.0 is based on a previous proposal developed by Team Athene UKP TU Darmstadt. Our proposal modifies the sentence retrieval phase, using statement extraction and representation in the form of triplets (subject, object, action). Triplets are extracted from the claim and compare to triplets extracted from Wikipedia articles using semantic similarity. Our results are satisfactory but there is room for improvement.","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126938033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Interactive Evidence Detection: train state-of-the-art model out-of-domain or simple model interactively? 交互式证据检测:交互式地训练最先进的域外模型还是简单模型?
C. Stahlhut
Finding evidence is of vital importance in research as well as fact checking and an evidence detection method would be useful in speeding up this process. However, when addressing a new topic there is no training data and there are two approaches to get started. One could use large amounts of out-of-domain data to train a state-of-the-art method, or to use the small data that a person creates while working on the topic. In this paper, we address this problem in two steps. First, by simulating users who read source documents and label sentences they can use as evidence, thereby creating small amounts of training data for an interactively trained evidence detection model; and second, by comparing such an interactively trained model against a pre-trained model that has been trained on large out-of-domain data. We found that an interactively trained model not only often out-performs a state-of-the-art model but also requires significantly lower amounts of computational resources. Therefore, especially when computational resources are scarce, e.g. no GPU available, training a smaller model on the fly is preferable to training a well generalising but resource hungry out-of-domain model.
寻找证据在研究和事实核查中都是至关重要的,证据检测方法将有助于加快这一进程。然而,当处理一个新主题时,没有训练数据,有两种方法可以开始。可以使用大量的域外数据来训练最先进的方法,或者使用一个人在研究该主题时创建的小数据。在本文中,我们分两步解决这个问题。首先,通过模拟用户阅读源文档并标记他们可以用作证据的句子,从而为交互式训练的证据检测模型创建少量训练数据;第二,通过将这种交互式训练模型与在大量域外数据上训练的预训练模型进行比较。我们发现,交互式训练模型不仅通常优于最先进的模型,而且需要的计算资源也大大减少。因此,特别是当计算资源稀缺时,例如没有可用的GPU,动态训练一个较小的模型比训练一个泛化良好但资源匮乏的域外模型更可取。
{"title":"Interactive Evidence Detection: train state-of-the-art model out-of-domain or simple model interactively?","authors":"C. Stahlhut","doi":"10.18653/v1/D19-6613","DOIUrl":"https://doi.org/10.18653/v1/D19-6613","url":null,"abstract":"Finding evidence is of vital importance in research as well as fact checking and an evidence detection method would be useful in speeding up this process. However, when addressing a new topic there is no training data and there are two approaches to get started. One could use large amounts of out-of-domain data to train a state-of-the-art method, or to use the small data that a person creates while working on the topic. In this paper, we address this problem in two steps. First, by simulating users who read source documents and label sentences they can use as evidence, thereby creating small amounts of training data for an interactively trained evidence detection model; and second, by comparing such an interactively trained model against a pre-trained model that has been trained on large out-of-domain data. We found that an interactively trained model not only often out-performs a state-of-the-art model but also requires significantly lower amounts of computational resources. Therefore, especially when computational resources are scarce, e.g. no GPU available, training a smaller model on the fly is preferable to training a well generalising but resource hungry out-of-domain model.","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133564047","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Neural Multi-Task Learning for Stance Prediction 姿态预测的神经多任务学习
Wei Fang, Moin Nadeem, Mitra Mohtarami, James R. Glass
We present a multi-task learning model that leverages large amount of textual information from existing datasets to improve stance prediction. In particular, we utilize multiple NLP tasks under both unsupervised and supervised settings for the target stance prediction task. Our model obtains state-of-the-art performance on a public benchmark dataset, Fake News Challenge, outperforming current approaches by a wide margin.
我们提出了一个多任务学习模型,利用现有数据集的大量文本信息来改进姿态预测。特别是,我们在目标姿态预测任务中使用了无监督和有监督设置下的多个NLP任务。我们的模型在公共基准数据集Fake News Challenge上获得了最先进的性能,远远优于当前的方法。
{"title":"Neural Multi-Task Learning for Stance Prediction","authors":"Wei Fang, Moin Nadeem, Mitra Mohtarami, James R. Glass","doi":"10.18653/v1/D19-6603","DOIUrl":"https://doi.org/10.18653/v1/D19-6603","url":null,"abstract":"We present a multi-task learning model that leverages large amount of textual information from existing datasets to improve stance prediction. In particular, we utilize multiple NLP tasks under both unsupervised and supervised settings for the target stance prediction task. Our model obtains state-of-the-art performance on a public benchmark dataset, Fake News Challenge, outperforming current approaches by a wide margin.","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115971030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Hybrid Models for Aspects Extraction without Labelled Dataset 无标记数据集的方面提取混合模型
W. Khong, Lay-Ki Soon, Hui-Ngo Goh
One of the important tasks in opinion mining is to extract aspects of the opinion target. Aspects are features or characteristics of the opinion target that are being reviewed, which can be categorised into explicit and implicit aspects. Extracting aspects from opinions is essential in order to ensure accurate information about certain attributes of an opinion target is retrieved. For instance, a professional camera receives a positive feedback in terms of its functionalities in a review, but its overly high price receives negative feedback. Most of the existing solutions focus on explicit aspects. However, sentences in reviews normally do not state the aspects explicitly. In this research, two hybrid models are proposed to identify and extract both explicit and implicit aspects, namely TDM-DC and TDM-TED. The proposed models combine topic modelling and dictionary-based approach. The models are unsupervised as they do not require any labelled dataset. The experimental results show that TDM-DC achieves F1-measure of 58.70%, where it outperforms both the baseline topic model and dictionary-based approach. In comparison to other existing unsupervised techniques, the proposed models are able to achieve higher F1-measure by approximately 3%. Although the supervised techniques perform slightly better, the proposed models are domain-independent, and hence more versatile.
意见挖掘的重要任务之一是提取意见目标的各个方面。方面是被审查的意见对象的特征或特征,可分为显性方面和隐性方面。为了确保检索到关于意见目标的某些属性的准确信息,从意见中提取方面是必不可少的。例如,一款专业相机在评测中就其功能得到了正面的反馈,但其过高的价格却得到了负面的反馈。大多数现有的解决方案都侧重于显式方面。然而,评论中的句子通常不会明确地陈述各个方面。本研究提出了两种混合模型,即TDM-DC和TDM-TED来识别和提取显式和隐式方面。提出的模型结合了主题建模和基于字典的方法。模型是无监督的,因为它们不需要任何标记的数据集。实验结果表明,TDM-DC的f1度量值达到58.70%,优于基线主题模型和基于字典的方法。与其他现有的无监督技术相比,所提出的模型能够实现大约3%的更高的f1测量。尽管有监督的技术表现稍好,但所提出的模型是领域无关的,因此更通用。
{"title":"Hybrid Models for Aspects Extraction without Labelled Dataset","authors":"W. Khong, Lay-Ki Soon, Hui-Ngo Goh","doi":"10.18653/v1/D19-6611","DOIUrl":"https://doi.org/10.18653/v1/D19-6611","url":null,"abstract":"One of the important tasks in opinion mining is to extract aspects of the opinion target. Aspects are features or characteristics of the opinion target that are being reviewed, which can be categorised into explicit and implicit aspects. Extracting aspects from opinions is essential in order to ensure accurate information about certain attributes of an opinion target is retrieved. For instance, a professional camera receives a positive feedback in terms of its functionalities in a review, but its overly high price receives negative feedback. Most of the existing solutions focus on explicit aspects. However, sentences in reviews normally do not state the aspects explicitly. In this research, two hybrid models are proposed to identify and extract both explicit and implicit aspects, namely TDM-DC and TDM-TED. The proposed models combine topic modelling and dictionary-based approach. The models are unsupervised as they do not require any labelled dataset. The experimental results show that TDM-DC achieves F1-measure of 58.70%, where it outperforms both the baseline topic model and dictionary-based approach. In comparison to other existing unsupervised techniques, the proposed models are able to achieve higher F1-measure by approximately 3%. Although the supervised techniques perform slightly better, the proposed models are domain-independent, and hence more versatile.","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"30 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120885272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
The FEVER2.0 Shared Task 共享任务
James Thorne, Andreas Vlachos, O. Cocarascu, Christos Christodoulopoulos, Arpit Mittal
We present the results of the second Fact Extraction and VERification (FEVER2.0) Shared Task. The task challenged participants to both build systems to verify factoid claims using evidence retrieved from Wikipedia and to generate adversarial attacks against other participant’s systems. The shared task had three phases: building, breaking and fixing. There were 8 systems in the builder’s round, three of which were new qualifying submissions for this shared task, and 5 adversaries generated instances designed to induce classification errors and one builder submitted a fixed system which had higher FEVER score and resilience than their first submission. All but one newly submitted systems attained FEVER scores higher than the best performing system from the first shared task and under adversarial evaluation, all systems exhibited losses in FEVER score. There was a great variety in adversarial attack types as well as the techniques used to generate the attacks, In this paper, we present the results of the shared task and a summary of the systems, highlighting commonalities and innovations among participating systems.
我们提出了第二个事实提取和验证(FEVER2.0)共享任务的结果。这项任务要求参与者既要构建系统,使用从维基百科检索到的证据来验证虚假声明,又要对其他参与者的系统产生对抗性攻击。共享任务分为三个阶段:构建、破坏和修复。在构建者的回合中有8个系统,其中3个是这个共享任务的新合格提交,5个对手生成了旨在诱导分类错误的实例,一个构建者提交了一个固定的系统,该系统比他们第一次提交的系统具有更高的FEVER分数和弹性。除了一个新提交的系统外,所有新提交的系统的FEVER得分都高于第一个共享任务中表现最好的系统,在对抗性评估下,所有系统的FEVER得分都有所下降。在本文中,我们展示了共享任务的结果和系统的总结,突出了参与系统之间的共性和创新。
{"title":"The FEVER2.0 Shared Task","authors":"James Thorne, Andreas Vlachos, O. Cocarascu, Christos Christodoulopoulos, Arpit Mittal","doi":"10.18653/v1/D19-6601","DOIUrl":"https://doi.org/10.18653/v1/D19-6601","url":null,"abstract":"We present the results of the second Fact Extraction and VERification (FEVER2.0) Shared Task. The task challenged participants to both build systems to verify factoid claims using evidence retrieved from Wikipedia and to generate adversarial attacks against other participant’s systems. The shared task had three phases: building, breaking and fixing. There were 8 systems in the builder’s round, three of which were new qualifying submissions for this shared task, and 5 adversaries generated instances designed to induce classification errors and one builder submitted a fixed system which had higher FEVER score and resilience than their first submission. All but one newly submitted systems attained FEVER scores higher than the best performing system from the first shared task and under adversarial evaluation, all systems exhibited losses in FEVER score. There was a great variety in adversarial attack types as well as the techniques used to generate the attacks, In this paper, we present the results of the shared task and a summary of the systems, highlighting commonalities and innovations among participating systems.","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129605811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 70
Scalable Knowledge Graph Construction from Text Collections 从文本集合构建可扩展的知识图谱
R. Clancy, I. Ilyas, Jimmy J. Lin
We present a scalable, open-source platform that “distills” a potentially large text collection into a knowledge graph. Our platform takes documents stored in Apache Solr and scales out the Stanford CoreNLP toolkit via Apache Spark integration to extract mentions and relations that are then ingested into the Neo4j graph database. The raw knowledge graph is then enriched with facts extracted from an external knowledge graph. The complete product can be manipulated by various applications using Neo4j’s native Cypher query language: We present a subgraph-matching approach to align extracted relations with external facts and show that fact verification, locating textual support for asserted facts, detecting inconsistent and missing facts, and extracting distantly-supervised training data can all be performed within the same framework.
我们提出了一个可扩展的开源平台,可以将潜在的大型文本集合“提炼”成知识图。我们的平台采用存储在Apache Solr中的文档,并通过Apache Spark集成扩展斯坦福CoreNLP工具包,以提取提及和关系,然后将其摄取到Neo4j图形数据库中。然后用从外部知识图中提取的事实来丰富原始知识图。完整的产品可以通过使用Neo4j的原生Cypher查询语言的各种应用程序来操作:我们提出了一种子图匹配方法,将提取的关系与外部事实对齐,并显示事实验证,定位断言事实的文本支持,检测不一致和缺失的事实,以及提取远程监督的训练数据都可以在同一个框架内执行。
{"title":"Scalable Knowledge Graph Construction from Text Collections","authors":"R. Clancy, I. Ilyas, Jimmy J. Lin","doi":"10.18653/v1/D19-6607","DOIUrl":"https://doi.org/10.18653/v1/D19-6607","url":null,"abstract":"We present a scalable, open-source platform that “distills” a potentially large text collection into a knowledge graph. Our platform takes documents stored in Apache Solr and scales out the Stanford CoreNLP toolkit via Apache Spark integration to extract mentions and relations that are then ingested into the Neo4j graph database. The raw knowledge graph is then enriched with facts extracted from an external knowledge graph. The complete product can be manipulated by various applications using Neo4j’s native Cypher query language: We present a subgraph-matching approach to align extracted relations with external facts and show that fact verification, locating textual support for asserted facts, detecting inconsistent and missing facts, and extracting distantly-supervised training data can all be performed within the same framework.","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"13 1-4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120963923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
FEVER Breaker’s Run of Team NbAuzDrLqg 狂热破坏者队NbAuzDrLqg的运行
Youngwoo Kim, J. Allan
We describe our submission for the Breaker phase of the second Fact Extraction and VERification (FEVER) Shared Task. Our adversarial data can be explained by two perspectives. First, we aimed at testing model’s ability to retrieve evidence, when appropriate query terms could not be easily generated from the claim. Second, we test model’s ability to precisely understand the implications of the texts, which we expect to be rare in FEVER 1.0 dataset. Overall, we suggested six types of adversarial attacks. The evaluation on the submitted systems showed that the systems were only able get both the evidence and label correct in 20% of the data. We also demonstrate our adversarial run analysis in the data development process.
我们描述了我们对第二个事实提取和验证(FEVER)共享任务的断路器阶段的提交。我们的对抗性数据可以用两个角度来解释。首先,我们的目标是测试模型检索证据的能力,当不容易从索赔中生成适当的查询条件时。其次,我们测试了模型精确理解文本含义的能力,我们预计这在FEVER 1.0数据集中是罕见的。总的来说,我们提出了六种类型的对抗性攻击。对提交系统的评价表明,系统仅能在20%的数据中同时获得证据和标签的正确性。我们还演示了数据开发过程中的对抗性运行分析。
{"title":"FEVER Breaker’s Run of Team NbAuzDrLqg","authors":"Youngwoo Kim, J. Allan","doi":"10.18653/v1/D19-6615","DOIUrl":"https://doi.org/10.18653/v1/D19-6615","url":null,"abstract":"We describe our submission for the Breaker phase of the second Fact Extraction and VERification (FEVER) Shared Task. Our adversarial data can be explained by two perspectives. First, we aimed at testing model’s ability to retrieve evidence, when appropriate query terms could not be easily generated from the claim. Second, we test model’s ability to precisely understand the implications of the texts, which we expect to be rare in FEVER 1.0 dataset. Overall, we suggested six types of adversarial attacks. The evaluation on the submitted systems showed that the systems were only able get both the evidence and label correct in 20% of the data. We also demonstrate our adversarial run analysis in the data development process.","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129297957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Fact Checking or Psycholinguistics: How to Distinguish Fake and True Claims? 事实检验或心理语言学:如何分辨真假?
A. Wawer, Grzegorz Wojdyga, Justyna Sarzyńska-Wawer
The goal of our paper is to compare psycholinguistic text features with fact checking approaches to distinguish lies from true statements. We examine both methods using data from a large ongoing study on deception and deception detection covering a mixture of factual and opinionated topics that polarize public opinion. We conclude that fact checking approaches based on Wikipedia are too limited for this task, as only a few percent of sentences from our study has enough evidence to become supported or refuted. Psycholinguistic features turn out to outperform both fact checking and human baselines, but the accuracy is not high. Overall, it appears that deception detection applicable to less-than-obvious topics is a difficult task and a problem to be solved.
本文的目的是比较心理语言学文本特征与事实检查方法,以区分谎言和真实陈述。我们使用一项正在进行的关于欺骗和欺骗检测的大型研究的数据来检验这两种方法,该研究涵盖了使公众舆论两极分化的事实和固执己见的主题。我们的结论是,基于维基百科的事实检查方法对于这项任务来说太有限了,因为我们研究中只有百分之几的句子有足够的证据来支持或反驳。心理语言学特征结果优于事实核查和人类基线,但准确性不高。总的来说,似乎欺骗检测适用于不太明显的主题是一项艰巨的任务和有待解决的问题。
{"title":"Fact Checking or Psycholinguistics: How to Distinguish Fake and True Claims?","authors":"A. Wawer, Grzegorz Wojdyga, Justyna Sarzyńska-Wawer","doi":"10.18653/v1/D19-6602","DOIUrl":"https://doi.org/10.18653/v1/D19-6602","url":null,"abstract":"The goal of our paper is to compare psycholinguistic text features with fact checking approaches to distinguish lies from true statements. We examine both methods using data from a large ongoing study on deception and deception detection covering a mixture of factual and opinionated topics that polarize public opinion. We conclude that fact checking approaches based on Wikipedia are too limited for this task, as only a few percent of sentences from our study has enough evidence to become supported or refuted. Psycholinguistic features turn out to outperform both fact checking and human baselines, but the accuracy is not high. Overall, it appears that deception detection applicable to less-than-obvious topics is a difficult task and a problem to be solved.","PeriodicalId":153447,"journal":{"name":"Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128352650","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1