首页 > 最新文献

Journal of Web Semantics最新文献

英文 中文
Time-aware evidence ranking for fact-checking 时间意识证据排序事实核查
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-11-01 DOI: 10.1016/j.websem.2021.100663
Liesbeth Allein , Isabelle Augenstein , Marie-Francine Moens

Truth can vary over time. Fact-checking decisions on claim veracity should therefore take into account temporal information of both the claim and supporting or refuting evidence. In this work, we investigate the hypothesis that the timestamp of a Web page is crucial to how it should be ranked for a given claim. We delineate four temporal ranking methods that constrain evidence ranking differently and simulate hypothesis-specific evidence rankings given the evidence timestamps as gold standard. Evidence ranking in three fact-checking models is ultimately optimized using a learning-to-rank loss function. Our study reveals that time-aware evidence ranking not only surpasses relevance assumptions based purely on semantic similarity or position in a search results list, but also improves veracity predictions of time-sensitive claims in particular.

真相会随着时间而变化。因此,关于索赔真实性的事实核查决定应考虑索赔和支持或反驳证据的时间信息。在这项工作中,我们研究了一个假设,即Web页面的时间戳对于如何对给定的索赔进行排名至关重要。我们描述了四种时间排序方法,以不同的方式约束证据排序,并在证据时间戳为金标准的情况下模拟特定于假设的证据排序。三种事实核查模型中的证据排序最终使用学习排序损失函数进行优化。我们的研究表明,时间敏感证据排序不仅超越了纯粹基于语义相似性或搜索结果列表中的位置的相关性假设,而且还提高了对时间敏感索赔的准确性预测。
{"title":"Time-aware evidence ranking for fact-checking","authors":"Liesbeth Allein ,&nbsp;Isabelle Augenstein ,&nbsp;Marie-Francine Moens","doi":"10.1016/j.websem.2021.100663","DOIUrl":"10.1016/j.websem.2021.100663","url":null,"abstract":"<div><p>Truth can vary over time. Fact-checking decisions on claim veracity should therefore take into account temporal information of both the claim and supporting or refuting evidence. In this work, we investigate the hypothesis that the timestamp of a Web page is crucial to how it should be ranked for a given claim. We delineate four temporal ranking methods that constrain evidence ranking differently and simulate hypothesis-specific evidence rankings given the evidence timestamps as gold standard. Evidence ranking in three fact-checking models is ultimately optimized using a learning-to-rank loss function. Our study reveals that time-aware evidence ranking not only surpasses relevance assumptions based purely on semantic similarity or position in a search results list, but also improves veracity predictions of time-sensitive claims in particular.</p></div>","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S157082682100038X/pdfft?md5=b9db24722ada9dc060b0028d53c7640b&pid=1-s2.0-S157082682100038X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87992041","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Optimizing a tableau reasoner and its implementation in Prolog 优化一个表推理器及其在Prolog中的实现
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-11-01 DOI: 10.1016/j.websem.2021.100677
Riccardo Zese , Giuseppe Cota

One of the foremost reasoning services for knowledge bases is finding all the justifications for a query. This is useful for debugging purpose and for coping with uncertainty. Among Description Logics (DLs) reasoners, the tableau algorithm is one of the most used. However, in order to collect the justifications, the reasoners must manage the non-determinism of the tableau method. For these reasons, a Prolog implementation can facilitate the management of such non-determinism.

The TRILL framework contains three probabilistic reasoners written in Prolog: TRILL, TRILLP and TORNADO. Since they are all part of the same framework, the choice about which to use can be done easily via the framework settings. Each one of them uses different approaches for probabilistic inference and handles different DLs flavors. Our previous work showed that they can sometimes achieve better results than state-of-the-art (non-)probabilistic reasoners.

In this paper we present two optimizations that improve the performances of the TRILL reasoners. The first one consists into identifying the fragment of the KB that allows to perform inference without losing the completeness. The second one modifies which tableau rule to apply and their order of application, in order to reduce the number of operations. Experimental results show the effectiveness of the introduced optimizations.

知识库最重要的推理服务之一是找到查询的所有理由。这对于调试和处理不确定性非常有用。在描述逻辑(dl)推理器中,表算法是最常用的一种。然而,为了收集理由,推理者必须管理表格方法的非决定论。由于这些原因,Prolog实现可以方便地管理这种不确定性。TRILL框架包含三个用Prolog编写的概率推理器:TRILL、TRILLP和TORNADO。由于它们都是同一框架的一部分,因此可以通过框架设置轻松地选择使用哪一个。它们中的每一个都使用不同的方法进行概率推理,并处理不同的dl风格。我们之前的工作表明,它们有时可以比最先进的(非)概率推理器获得更好的结果。在本文中,我们提出了两种改进TRILL推理器性能的优化方法。第一个步骤包括识别允许在不丢失完整性的情况下执行推理的知识库片段。第二种方法修改要应用的表规则及其应用顺序,以减少操作的数量。实验结果表明了所引入的优化方法的有效性。
{"title":"Optimizing a tableau reasoner and its implementation in Prolog","authors":"Riccardo Zese ,&nbsp;Giuseppe Cota","doi":"10.1016/j.websem.2021.100677","DOIUrl":"https://doi.org/10.1016/j.websem.2021.100677","url":null,"abstract":"<div><p><span>One of the foremost reasoning services for knowledge bases is finding all the justifications for a query. This is useful for debugging purpose and for coping with uncertainty. Among </span>Description Logics<span><span> (DLs) reasoners, the tableau algorithm is one of the most used. However, in order to collect the justifications, the reasoners must manage the non-determinism of the </span>tableau method. For these reasons, a Prolog implementation can facilitate the management of such non-determinism.</span></p><p>The TRILL framework contains three probabilistic reasoners written in Prolog: TRILL, TRILL<sup><em>P</em></sup><span> and TORNADO. Since they are all part of the same framework, the choice about which to use can be done easily via the framework settings. Each one of them uses different approaches for probabilistic inference and handles different DLs flavors. Our previous work showed that they can sometimes achieve better results than state-of-the-art (non-)probabilistic reasoners.</span></p><p>In this paper we present two optimizations that improve the performances of the TRILL reasoners. The first one consists into identifying the fragment of the KB that allows to perform inference without losing the completeness. The second one modifies which tableau rule to apply and their order of application, in order to reduce the number of operations. Experimental results show the effectiveness of the introduced optimizations.</p></div>","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138090273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards the next generation of the LinkedGeoData project using virtual knowledge graphs 迈向下一代使用虚拟知识图谱的LinkedGeoData项目
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-11-01 DOI: 10.1016/j.websem.2021.100662
Linfang Ding , Guohui Xiao , Albulen Pano , Claus Stadler , Diego Calvanese

With the advancement of Semantic Technologies, large geospatial data sources have been increasingly published as Linked data on the Web. The LinkedGeoData project is one of the most prominent such projects to create a large knowledge graph from OpenStreetMap (OSM) with global coverage and interlinking of other data sources. In this paper, we report on the ongoing effort of exposing the relational database in LinkedGeoData as a SPARQL endpoint using Virtual Knowledge Graph (VKG) technology. Specifically, we present two realizations of VKGs, using the two systems Sparqlify and Ontop. In order to improve compliance with the OGC GeoSPARQL standard, we have implemented GeoSPARQL support in Ontop v4. Moreover, we have evaluated the VKG-powered LinkedGeoData in the test areas of Italy and Germany. Our experiments demonstrate that such system supports complex GeoSPARQL queries, which confirms that query answering in the VKG approach is efficient.

随着语义技术的发展,越来越多的大型地理空间数据源以链接数据的形式在Web上发布。LinkedGeoData项目是此类项目中最突出的一个,它从OpenStreetMap (OSM)中创建一个具有全球覆盖范围和其他数据源相互链接的大型知识图谱。在本文中,我们报告了使用虚拟知识图(VKG)技术将LinkedGeoData中的关系数据库作为SPARQL端点公开的正在进行的工作。具体来说,我们介绍了两种vkg的实现,使用Sparqlify和Ontop两个系统。为了改进对OGC GeoSPARQL标准的遵从性,我们在Ontop v4中实现了GeoSPARQL支持。此外,我们在意大利和德国的测试地区评估了vkg驱动的LinkedGeoData。实验表明,该系统支持复杂的GeoSPARQL查询,这证实了VKG方法的查询应答是有效的。
{"title":"Towards the next generation of the LinkedGeoData project using virtual knowledge graphs","authors":"Linfang Ding ,&nbsp;Guohui Xiao ,&nbsp;Albulen Pano ,&nbsp;Claus Stadler ,&nbsp;Diego Calvanese","doi":"10.1016/j.websem.2021.100662","DOIUrl":"10.1016/j.websem.2021.100662","url":null,"abstract":"<div><p>With the advancement of Semantic Technologies, large geospatial data sources have been increasingly published as Linked data on the Web. The <em>LinkedGeoData</em> project is one of the most prominent such projects to create a large knowledge graph from <em>OpenStreetMap</em> (OSM) with global coverage and interlinking of other data sources. In this paper, we report on the ongoing effort of exposing the relational database in LinkedGeoData as a SPARQL endpoint using <em>Virtual Knowledge Graph</em> (VKG) technology. Specifically, we present two realizations of VKGs, using the two systems Sparqlify and Ontop. In order to improve compliance with the OGC GeoSPARQL standard, we have implemented GeoSPARQL support in Ontop v4. Moreover, we have evaluated the VKG-powered LinkedGeoData in the test areas of Italy and Germany. Our experiments demonstrate that such system supports complex GeoSPARQL queries, which confirms that query answering in the VKG approach is efficient.</p></div>","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1570826821000378/pdfft?md5=4faeffc9e304792fcc147abfd1e837e1&pid=1-s2.0-S1570826821000378-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89708603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Design, Application and Evaluation of PROV-SwProcess: A PROV extension Data Model for Software Development Processes PROV- swprocess的设计、应用与评价:软件开发过程的PROV扩展数据模型
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-11-01 DOI: 10.1016/j.websem.2021.100676
Gabriella Castro Barbosa Costa , Claudia Werner , Regina Braga , Eldânae Nogueira Teixeira , Victor Ströele , Marco Antônio Pereira Araújo , Marcos Alexandre Miguel

The literature defines data provenance as the description of the origins of a piece of data and the process by which it arrived in a database. It helps to audit and understand data history and bring transparency to the process. Provenance has been successfully used in scientific computing, chemical industries, and health sciences, considering that these areas require a comprehensive traceability mechanism. Meanwhile, companies have been collecting and storing more data from their systems and processes. This work investigates if the use of provenance models and techniques can support software processes execution analysis and data-driven decision-making, considering the increasing availability of process data provided by companies. PROV-SwProcess, a software development process provenance modeling proposal, was developed and evaluated by process and provenance experts. Our proposal is an extension of the W3C recommended standard model PROV, aiming to capture and store the most relevant information about software development process provenance data. The results suggest that the model’s suitability improves and assists process managers in the software process analysis and supports data-driven decision-making.

文献将数据来源定义为对一段数据的起源及其到达数据库的过程的描述。它有助于审计和理解数据历史,并为流程带来透明度。考虑到这些领域需要全面的可追溯性机制,来源已经成功地应用于科学计算、化学工业和健康科学。与此同时,公司一直在从他们的系统和流程中收集和存储更多的数据。这项工作调查了来源模型和技术的使用是否能够支持软件过程执行分析和数据驱动的决策,考虑到公司提供的过程数据的日益增加的可用性。PROV-SwProcess是一个软件开发过程来源建模建议,由过程和来源专家开发和评估。我们的建议是W3C推荐的标准模型PROV的扩展,旨在捕获和存储关于软件开发过程起源数据的最相关信息。结果表明,该模型的适用性提高并有助于过程管理者进行软件过程分析,支持数据驱动的决策。
{"title":"Design, Application and Evaluation of PROV-SwProcess: A PROV extension Data Model for Software Development Processes","authors":"Gabriella Castro Barbosa Costa ,&nbsp;Claudia Werner ,&nbsp;Regina Braga ,&nbsp;Eldânae Nogueira Teixeira ,&nbsp;Victor Ströele ,&nbsp;Marco Antônio Pereira Araújo ,&nbsp;Marcos Alexandre Miguel","doi":"10.1016/j.websem.2021.100676","DOIUrl":"10.1016/j.websem.2021.100676","url":null,"abstract":"<div><p>The literature defines data provenance<span> as the description of the origins of a piece of data and the process by which it arrived in a database. It helps to audit and understand data history and bring transparency to the process. Provenance has been successfully used in scientific computing, chemical industries, and health sciences, considering that these areas require a comprehensive traceability mechanism. Meanwhile, companies have been collecting and storing more data from their systems and processes. This work investigates if the use of provenance models and techniques can support software processes execution analysis and data-driven decision-making, considering the increasing availability of process data provided by companies. PROV-SwProcess, a software development process provenance modeling proposal, was developed and evaluated by process and provenance experts. Our proposal is an extension of the W3C recommended standard model PROV, aiming to capture and store the most relevant information about software development process provenance data. The results suggest that the model’s suitability improves and assists process managers in the software process analysis and supports data-driven decision-making.</span></p></div>","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77723070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
SemML: Facilitating development of ML models for condition monitoring with semantics SemML:促进使用语义进行状态监测的ML模型的开发
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-11-01 DOI: 10.1016/j.websem.2021.100664
Baifan Zhou , Yulia Svetashova , Andre Gusmao , Ahmet Soylu , Gong Cheng , Ralf Mikut , Arild Waaler , Evgeny Kharlamov

Monitoring of the state, performance, quality of operations and other parameters of equipment and production processes, which is typically referred to as condition monitoring, is an important common practice in many industries including manufacturing, oil and gas, chemical and process industry. In the age of Industry 4.0, where the aim is a deep degree of production automation, unprecedented amounts of data are generated by equipment and processes, and this enables adoption of Machine Learning (ML) approaches for condition monitoring. Development of such ML models is challenging. On the one hand, it requires collaborative work of experts from different areas, including data scientists, engineers, process experts, and managers with asymmetric backgrounds. On the other hand, there is high variety and diversity of data relevant for condition monitoring. Both factors hampers ML modelling for condition monitoring. In this work, we address these challenges by empowering ML-based condition monitoring with semantic technologies. To this end we propose a software system SemML that allows to reuse and generalise ML pipelines for conditions monitoring by relying on semantics. In particular, SemML has several novel components and relies on ontologies and ontology templates for ML task negotiation and for data and ML feature annotation. SemML also allows to instantiate parametrised ML pipelines by semantic annotation of industrial data. With SemML, users do not need to dive into data and ML scripts when new datasets of a studied application scenario arrive. They only need to annotate data and then ML models will be constructed through the combination of semantic reasoning and ML modules. We demonstrate the benefits of SemML on a Bosch use-case of electric resistance welding with very promising results.

对设备和生产过程的状态、性能、运行质量和其他参数进行监测,通常称为状态监测,是包括制造业、石油和天然气、化工和加工工业在内的许多行业的重要惯例。在工业4.0时代,其目标是实现深度生产自动化,设备和工艺产生了前所未有的数据量,这使得采用机器学习(ML)方法进行状态监测成为可能。这种机器学习模型的开发具有挑战性。一方面,它需要来自不同领域的专家的协同工作,包括数据科学家、工程师、流程专家和背景不对称的管理人员。另一方面,与状态监测相关的数据种类繁多。这两个因素都阻碍了状态监测的ML建模。在这项工作中,我们通过使用语义技术增强基于机器学习的状态监测来解决这些挑战。为此,我们提出了一个软件系统SemML,它允许通过依赖语义来重用和泛化ML管道进行状态监控。特别地,SemML有几个新颖的组件,并且依赖于本体和本体模板来进行ML任务协商以及数据和ML特性注释。SemML还允许通过对工业数据的语义注释来实例化参数化的ML管道。使用SemML,当所研究的应用程序场景的新数据集到达时,用户不需要深入研究数据和ML脚本。他们只需要对数据进行标注,然后通过语义推理和ML模块的结合来构建ML模型。我们在博世的电阻焊接用例中展示了SemML的优势,并取得了非常有希望的结果。
{"title":"SemML: Facilitating development of ML models for condition monitoring with semantics","authors":"Baifan Zhou ,&nbsp;Yulia Svetashova ,&nbsp;Andre Gusmao ,&nbsp;Ahmet Soylu ,&nbsp;Gong Cheng ,&nbsp;Ralf Mikut ,&nbsp;Arild Waaler ,&nbsp;Evgeny Kharlamov","doi":"10.1016/j.websem.2021.100664","DOIUrl":"10.1016/j.websem.2021.100664","url":null,"abstract":"<div><p>Monitoring of the state, performance, quality of operations and other parameters of equipment and production processes, which is typically referred to as condition monitoring, is an important common practice in many industries including manufacturing, oil and gas, chemical and process industry. In the age of Industry 4.0, where the aim is a deep degree of production automation, unprecedented amounts of data are generated by equipment and processes, and this enables adoption of Machine Learning (ML) approaches for condition monitoring. Development of such ML models is challenging. On the one hand, it requires collaborative work of experts from different areas, including data scientists, engineers, process experts, and managers with asymmetric backgrounds. On the other hand, there is high variety and diversity of data relevant for condition monitoring. Both factors hampers ML modelling for condition monitoring. In this work, we address these challenges by empowering ML-based condition monitoring with semantic technologies. To this end we propose a software system SemML that allows to reuse and generalise ML pipelines for conditions monitoring by relying on semantics. In particular, SemML has several novel components and relies on ontologies and ontology templates for ML task negotiation and for data and ML feature annotation. SemML also allows to instantiate parametrised ML pipelines by semantic annotation of industrial data. With SemML, users do not need to dive into data and ML scripts when new datasets of a studied application scenario arrive. They only need to annotate data and then ML models will be constructed through the combination of semantic reasoning and ML modules. We demonstrate the benefits of SemML on a Bosch use-case of electric resistance welding with very promising results.</p></div>","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1570826821000391/pdfft?md5=13aa3514aea684bfe3a850fdc768a1e0&pid=1-s2.0-S1570826821000391-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89704076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Negative statements considered useful 被认为有用的否定陈述
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-11-01 DOI: 10.1016/j.websem.2021.100661
Hiba Arnaout , Simon Razniewski , Gerhard Weikum , Jeff Z. Pan

Knowledge bases (KBs) about notable entities and their properties are an important asset in applications such as search, question answering and dialog. All popular KBs capture virtually only positive statements, and abstain from taking any stance on statements not stored in the KB. This paper makes the case for explicitly stating salient statements that do not hold. Negative statements are useful to overcome limitations of question answering systems that are mainly geared for positive questions; they can also contribute to informative summaries of entities. Due to the abundance of such invalid statements, any effort to compile them needs to address ranking by saliency. We present a statistical inference method for compiling and ranking negative statements, based on expectations from positive statements of related entities in peer groups. Experimental results, with a variety of datasets, show that the method can effectively discover notable negative statements, and extrinsic studies underline their usefulness for entity summarization. Datasets and code are released as resources for further research.

关于显著实体及其属性的知识库(KBs)是搜索、问答和对话等应用程序的重要资产。所有流行的KB实际上都只捕获正面语句,并且避免对未存储在KB中的语句采取任何立场。本文提出了明确陈述不成立的重要陈述的理由。否定陈述句有助于克服主要针对肯定问题的问答系统的局限性;它们还有助于提供实体的信息摘要。由于此类无效语句大量存在,因此编译它们的任何努力都需要解决显著性排序问题。我们提出了一种统计推理方法,用于编制和排序负面陈述,基于对同行群体中相关实体的积极陈述的期望。使用多种数据集的实验结果表明,该方法可以有效地发现显著的否定陈述,并且外部研究强调了其对实体总结的有用性。数据集和代码作为进一步研究的资源发布。
{"title":"Negative statements considered useful","authors":"Hiba Arnaout ,&nbsp;Simon Razniewski ,&nbsp;Gerhard Weikum ,&nbsp;Jeff Z. Pan","doi":"10.1016/j.websem.2021.100661","DOIUrl":"10.1016/j.websem.2021.100661","url":null,"abstract":"<div><p>Knowledge bases (KBs) about notable entities and their properties are an important asset in applications such as search, question answering and dialog. All popular KBs capture virtually only positive statements, and abstain from taking any stance on statements not stored in the KB. This paper makes the case for explicitly stating salient statements that do <em>not</em><span> hold. Negative statements are useful to overcome limitations of question answering systems that are mainly geared for positive questions; they can also contribute to informative summaries of entities. Due to the abundance of such invalid statements, any effort to compile them needs to address ranking by saliency. We present a statistical inference method for compiling and ranking negative statements, based on expectations from positive statements of related entities in peer groups. Experimental results, with a variety of datasets, show that the method can effectively discover notable negative statements, and extrinsic studies underline their usefulness for entity summarization. Datasets and code are released as resources for further research.</span></p></div>","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89778494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Intelligent software web agents: A gap analysis 智能软件网络代理:差距分析
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-11-01 DOI: 10.1016/j.websem.2021.100659
Sabrina Kirrane

Semantic web technologies have shown their effectiveness, especially when it comes to knowledge representation, reasoning, and data integration. However, the original semantic web vision, whereby machine readable web data could be automatically actioned upon by intelligent software web agents, has yet to be realised. In order to better understand the existing technological opportunities and challenges, in this paper we examine the status quo in terms of intelligent software web agents, guided by research with respect to requirements and architectural components, coming from the agents community. We use the identified requirements to both further elaborate on the semantic web agent motivating use case scenario, and to summarise different perspectives on the requirements from the semantic web agent literature. We subsequently propose a hybrid semantic web agent architecture, and use the various components and subcomponents in order to provide a focused discussion in relation to existing semantic web standards and community activities. Finally, we highlight open research opportunities and challenges and take a broader perspective of the research by discussing the potential for intelligent software web agents as an enabling technology for emerging domains, such as digital assistants, cloud computing, and the internet of things.

语义网技术已经显示出其有效性,特别是在知识表示、推理和数据集成方面。然而,原始的语义网络视觉,即机器可读的网络数据可以由智能软件网络代理自动操作,尚未实现。为了更好地理解现有的技术机遇和挑战,在本文中,我们通过对来自代理社区的需求和体系结构组件的研究来研究智能软件web代理的现状。我们使用确定的需求来进一步阐述语义web代理激励用例场景,并总结语义web代理文献中对需求的不同观点。我们随后提出了一个混合语义web代理架构,并使用各种组件和子组件,以便提供与现有语义web标准和社区活动相关的重点讨论。最后,我们强调了开放的研究机会和挑战,并通过讨论智能软件网络代理作为新兴领域(如数字助理、云计算和物联网)的使能技术的潜力,以更广阔的视角进行了研究。
{"title":"Intelligent software web agents: A gap analysis","authors":"Sabrina Kirrane","doi":"10.1016/j.websem.2021.100659","DOIUrl":"10.1016/j.websem.2021.100659","url":null,"abstract":"<div><p>Semantic web technologies have shown their effectiveness, especially when it comes to knowledge representation, reasoning, and data integration. However, the original semantic web vision, whereby machine readable web data could be automatically actioned upon by intelligent software web agents, has yet to be realised. In order to better understand the existing technological opportunities and challenges, in this paper we examine the status quo in terms of intelligent software web agents, guided by research with respect to requirements and architectural components, coming from the agents community. We use the identified requirements to both further elaborate on the semantic web agent motivating use case scenario, and to summarise different perspectives on the requirements from the semantic web agent literature. We subsequently propose a hybrid semantic web agent architecture, and use the various components and subcomponents in order to provide a focused discussion in relation to existing semantic web standards and community activities. Finally, we highlight open research opportunities and challenges and take a broader perspective of the research by discussing the potential for intelligent software web agents as an enabling technology for emerging domains, such as digital assistants, cloud computing, and the internet of things.</p></div>","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1570826821000342/pdfft?md5=5e359d8b7b907172f9f83e621c1e421a&pid=1-s2.0-S1570826821000342-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88079819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
HeadlineStanceChecker: Exploiting summarization to detect headline disinformation HeadlineStanceChecker:利用摘要来检测标题虚假信息
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-11-01 DOI: 10.1016/j.websem.2021.100660
Robiert Sepúlveda-Torres, Marta Vicente, Estela Saquete, Elena Lloret, Manuel Palomar

The headline of a news article is designed to succinctly summarize its content, providing the reader with a clear understanding of the news item. Unfortunately, in the post-truth era, headlines are more focused on attracting the reader’s attention for ideological or commercial reasons, thus leading to mis- or disinformation through false or distorted headlines. One way of combating this, although a challenging task, is by determining the relation between the headline and the body text to establish the stance. Hence, to contribute to the detection of mis- and disinformation, this paper proposes an approach (HeadlineStanceChecker) that determines the stance of a headline with respect to the body text to which it is associated. The novelty rests on the use of a two-stage classification architecture that uses summarization techniques to shape the input for both classifiers instead of directly passing the full news body text, thereby reducing the amount of information to be processed while keeping important information. Specifically, summarization is done through Positional Language Models leveraging on semantic resources to identify salient information in the body text that is then compared to its corresponding headline. The results obtained show that our approach achieves 94.31% accuracy for the overall classification and the best FNC-1 relative score compared with the state of the art. It is especially remarkable that the system, which uses only the relevant information provided by the automatic summaries instead of the whole text, is able to classify the different stance categories with very competitive results, especially in the discuss stance between the headline and the news body text. It can be concluded that using automatic extractive summaries as input of our approach together with the two-stage architecture is an appropriate solution to the problem.

新闻文章的标题旨在简洁地概括其内容,使读者对新闻项目有一个清晰的了解。不幸的是,在后真相时代,新闻标题更多的是出于意识形态或商业原因而吸引读者的注意力,从而通过虚假或扭曲的标题产生错误或虚假的信息。解决这个问题的一种方法是确定标题和正文之间的关系,以确立立场,尽管这是一项具有挑战性的任务。因此,为了有助于检测错误和虚假信息,本文提出了一种方法(HeadlineStanceChecker),该方法确定标题相对于与其相关的正文文本的立场。其新颖之处在于使用了两阶段分类架构,该架构使用摘要技术为两个分类器塑造输入,而不是直接传递完整的新闻正文,从而减少了需要处理的信息量,同时保留了重要的信息。具体来说,摘要是通过位置语言模型利用语义资源来识别正文中的重要信息,然后将其与相应的标题进行比较。结果表明,我们的方法在整体分类上达到了94.31%的准确率,并且在FNC-1相对得分上达到了最佳水平。特别值得注意的是,该系统仅使用自动摘要提供的相关信息,而不是全文,能够对不同的立场类别进行分类,结果非常有竞争力,特别是在标题和新闻正文之间的讨论立场。可以得出结论,使用自动提取摘要作为我们方法的输入,并结合两阶段体系结构是解决问题的合适方案。
{"title":"HeadlineStanceChecker: Exploiting summarization to detect headline disinformation","authors":"Robiert Sepúlveda-Torres,&nbsp;Marta Vicente,&nbsp;Estela Saquete,&nbsp;Elena Lloret,&nbsp;Manuel Palomar","doi":"10.1016/j.websem.2021.100660","DOIUrl":"10.1016/j.websem.2021.100660","url":null,"abstract":"<div><p>The headline of a news article is designed to succinctly summarize its content, providing the reader with a clear understanding of the news item. Unfortunately, in the post-truth era, headlines are more focused on attracting the reader’s attention for ideological or commercial reasons, thus leading to mis- or disinformation through false or distorted headlines. One way of combating this, although a challenging task, is by determining the relation between the headline and the body text to establish the stance. Hence, to contribute to the detection of mis- and disinformation, this paper proposes an approach (<em>HeadlineStanceChecker</em>) that determines the stance of a headline with respect to the body text to which it is associated. The novelty rests on the use of a two-stage classification architecture that uses summarization techniques to shape the input for both classifiers instead of directly passing the full news body text, thereby reducing the amount of information to be processed while keeping important information. Specifically, summarization is done through Positional Language Models leveraging on semantic resources to identify salient information in the body text that is then compared to its corresponding headline. The results obtained show that our approach achieves 94.31% accuracy for the overall classification and the best FNC-1 relative score compared with the state of the art. It is especially remarkable that the system, which uses only the relevant information provided by the automatic summaries instead of the whole text, is able to classify the different stance categories with very competitive results, especially in the <em>discuss</em> stance between the headline and the news body text. It can be concluded that using automatic extractive summaries as input of our approach together with the two-stage architecture is an appropriate solution to the problem.</p></div>","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1570826821000354/pdfft?md5=2e9f623b6b4a0278d46a5df6af6c5671&pid=1-s2.0-S1570826821000354-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75634742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Optimizing a tableau reasoner and its implementation in Prolog 优化一个表推理器及其在Prolog中的实现
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-10-01 DOI: 10.2139/ssrn.3945445
Riccardo Zese, Giuseppe Cota
{"title":"Optimizing a tableau reasoner and its implementation in Prolog","authors":"Riccardo Zese, Giuseppe Cota","doi":"10.2139/ssrn.3945445","DOIUrl":"https://doi.org/10.2139/ssrn.3945445","url":null,"abstract":"","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86499553","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Crossing the chasm between ontology engineering and application development: A survey 跨越本体工程与应用程序开发之间的鸿沟:综述
IF 2.5 3区 计算机科学 Q1 Computer Science Pub Date : 2021-07-01 DOI: 10.1016/j.websem.2021.100655
Paola Espinoza-Arias , Daniel Garijo , Oscar Corcho

The adoption of Knowledge Graphs (KGs) by public and private organizations to integrate and publish data has increased in recent years. Ontologies play a crucial role in providing the structure for KGs, but are usually disregarded when designing Application Programming Interfaces (APIs) to enable browsing KGs in a developer-friendly manner. In this paper we provide a systematic review of the state of the art on existing approaches to ease access to ontology-based KG data by application developers. We propose two comparison frameworks to understand specifications, technologies and tools responsible for providing APIs for KGs. Our results reveal several limitations on existing API-based specifications, technologies and tools for KG consumption, which outline exciting research challenges including automatic API generation, API resource path prediction, ontology-based API versioning, and API validation and testing.

近年来,公共和私人组织越来越多地采用知识图谱(Knowledge Graphs, KGs)来整合和发布数据。本体在为KGs提供结构方面起着至关重要的作用,但在设计应用程序编程接口(api)以支持以开发人员友好的方式浏览KGs时,通常会忽略本体。在本文中,我们系统地回顾了现有方法的现状,以方便应用程序开发人员访问基于本体的KG数据。我们提出了两个比较框架来理解负责为KG提供API的规范、技术和工具。我们的研究结果揭示了现有的基于API的KG消费规范、技术和工具的一些局限性,其中概述了令人兴奋的研究挑战,包括自动API生成、API资源路径预测、基于本体的API版本控制以及API验证和测试。
{"title":"Crossing the chasm between ontology engineering and application development: A survey","authors":"Paola Espinoza-Arias ,&nbsp;Daniel Garijo ,&nbsp;Oscar Corcho","doi":"10.1016/j.websem.2021.100655","DOIUrl":"10.1016/j.websem.2021.100655","url":null,"abstract":"<div><p>The adoption of Knowledge Graphs (KGs) by public and private organizations to integrate and publish data has increased in recent years. Ontologies play a crucial role in providing the structure for KGs, but are usually disregarded when designing Application Programming Interfaces (APIs) to enable browsing KGs in a developer-friendly manner. In this paper we provide a systematic review of the state of the art on existing approaches to ease access to ontology-based KG data by application developers. We propose two comparison frameworks to understand specifications, technologies and tools responsible for providing APIs for KGs. Our results reveal several limitations on existing API-based specifications, technologies and tools for KG consumption, which outline exciting research challenges including automatic API generation, API resource path prediction, ontology-based API versioning, and API validation and testing.</p></div>","PeriodicalId":49951,"journal":{"name":"Journal of Web Semantics","volume":null,"pages":null},"PeriodicalIF":2.5,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.websem.2021.100655","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87299102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
Journal of Web Semantics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1