首页 > 最新文献

Semantic Web最新文献

英文 中文
A neuro-symbolic system over knowledge graphs for link prediction 基于知识图的链接预测神经符号系统
IF 3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-06-07 DOI: 10.3233/sw-233324
Ariam Rivas, D. Collarana, M. Torrente, Maria-Esther Vidal
Neuro-Symbolic Artificial Intelligence (AI) focuses on integrating symbolic and sub-symbolic systems to enhance the performance and explainability of predictive models. Symbolic and sub-symbolic approaches differ fundamentally in how they represent data and make use of data features to reach conclusions. Neuro-symbolic systems have recently received significant attention in the scientific community. However, despite efforts in neural-symbolic integration, symbolic processing can still be better exploited, mainly when these hybrid approaches are defined on top of knowledge graphs. This work is built on the statement that knowledge graphs can naturally represent the convergence between data and their contextual meaning (i.e., knowledge). We propose a hybrid system that resorts to symbolic reasoning, expressed as a deductive database, to augment the contextual meaning of entities in a knowledge graph, thus, improving the performance of link prediction implemented using knowledge graph embedding (KGE) models. An entity context is defined as the ego network of the entity in a knowledge graph. Given a link prediction task, the proposed approach deduces new RDF triples in the ego networks of the entities corresponding to the heads and tails of the prediction task on the knowledge graph (KG). Since knowledge graphs may be incomplete and sparse, the facts deduced by the symbolic system not only reduce sparsity but also make explicit meaningful relations among the entities that compose an entity ego network. As a proof of concept, our approach is applied over a KG for lung cancer to predict treatment effectiveness. The empirical results put the deduction power of deductive databases into perspective. They indicate that making explicit deduced relationships in the ego networks empowers all the studied KGE models to generate more accurate links.
神经符号人工智能(AI)专注于整合符号和子符号系统,以提高预测模型的性能和可解释性。符号方法和子符号方法在如何表示数据和利用数据特征得出结论方面存在根本的不同。神经符号系统最近在科学界受到了极大的关注。然而,尽管在神经-符号整合方面做出了努力,符号处理仍然可以更好地利用,主要是当这些混合方法被定义在知识图之上时。这项工作是建立在知识图可以自然地表示数据与其上下文含义(即知识)之间的融合的声明之上的。我们提出了一种混合系统,该系统采用符号推理,表示为演绎数据库,以增强知识图中实体的上下文含义,从而提高使用知识图嵌入(KGE)模型实现的链接预测的性能。实体上下文被定义为知识图谱中实体的自我网络。给定一个链接预测任务,该方法在知识图(KG)上预测任务的正反面对应的实体的自我网络中推导出新的RDF三元组。由于知识图可能是不完整的和稀疏的,通过符号系统推导出的事实不仅降低了稀疏性,而且在构成实体自我网络的实体之间建立了明确的有意义的关系。作为概念验证,我们的方法应用于肺癌的KG来预测治疗效果。实证结果揭示了演绎数据库的演绎能力。他们指出,在自我网络中建立明确的推导关系,使所有研究的KGE模型都能产生更准确的联系。
{"title":"A neuro-symbolic system over knowledge graphs for link prediction","authors":"Ariam Rivas, D. Collarana, M. Torrente, Maria-Esther Vidal","doi":"10.3233/sw-233324","DOIUrl":"https://doi.org/10.3233/sw-233324","url":null,"abstract":"Neuro-Symbolic Artificial Intelligence (AI) focuses on integrating symbolic and sub-symbolic systems to enhance the performance and explainability of predictive models. Symbolic and sub-symbolic approaches differ fundamentally in how they represent data and make use of data features to reach conclusions. Neuro-symbolic systems have recently received significant attention in the scientific community. However, despite efforts in neural-symbolic integration, symbolic processing can still be better exploited, mainly when these hybrid approaches are defined on top of knowledge graphs. This work is built on the statement that knowledge graphs can naturally represent the convergence between data and their contextual meaning (i.e., knowledge). We propose a hybrid system that resorts to symbolic reasoning, expressed as a deductive database, to augment the contextual meaning of entities in a knowledge graph, thus, improving the performance of link prediction implemented using knowledge graph embedding (KGE) models. An entity context is defined as the ego network of the entity in a knowledge graph. Given a link prediction task, the proposed approach deduces new RDF triples in the ego networks of the entities corresponding to the heads and tails of the prediction task on the knowledge graph (KG). Since knowledge graphs may be incomplete and sparse, the facts deduced by the symbolic system not only reduce sparsity but also make explicit meaningful relations among the entities that compose an entity ego network. As a proof of concept, our approach is applied over a KG for lung cancer to predict treatment effectiveness. The empirical results put the deduction power of deductive databases into perspective. They indicate that making explicit deduced relationships in the ego networks empowers all the studied KGE models to generate more accurate links.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"51 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2023-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79133289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
MuHeQA: Zero-shot question answering over multiple and heterogeneous knowledge bases MuHeQA:基于多个异构知识库的零概率问答
IF 3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-06-07 DOI: 10.3233/sw-233379
Carlos Badenes-Olmedo, Óscar Corcho
There are two main limitations in most of the existing Knowledge Graph Question Answering (KGQA) algorithms. First, the approaches depend heavily on the structure and cannot be easily adapted to other KGs. Second, the availability and amount of additional domain-specific data in structured or unstructured formats has also proven to be critical in many of these systems. Such dependencies limit the applicability of KGQA systems and make their adoption difficult. A novel algorithm is proposed, MuHeQA, that alleviates both limitations by retrieving the answer from textual content automatically generated from KGs instead of queries over them. This new approach (1) works on one or several KGs simultaneously, (2) does not require training data what makes it is domain-independent, (3) enables the combination of knowledge graphs with unstructured information sources to build the answer, and (4) reduces the dependency on the underlying schema since it does not navigate through structured content but only reads property values. MuHeQA extracts answers from textual summaries created by combining information related to the question from multiple knowledge bases, be them structured or not. Experiments over Wikidata and DBpedia show that our approach achieves comparable performance to other approaches in single-fact questions while being domain and KG independent. Results raise important questions for future work about how the textual content that can be created from knowledge graphs enables answer extraction.
大多数现有的知识图谱问答(KGQA)算法存在两个主要的局限性。首先,这些方法严重依赖于结构,不容易适应其他kg。其次,在许多这些系统中,结构化或非结构化格式的额外领域特定数据的可用性和数量也被证明是至关重要的。这种依赖限制了KGQA系统的适用性,并使其难以采用。提出了一种新的算法MuHeQA,通过从KGs自动生成的文本内容中检索答案,而不是对它们进行查询,从而减轻了这两种限制。这种新方法(1)同时在一个或多个KGs上工作,(2)不需要训练数据,这使得它是领域独立的,(3)允许知识图与非结构化信息源的组合来构建答案,(4)减少了对底层模式的依赖,因为它不浏览结构化内容,而只读取属性值。MuHeQA从文本摘要中提取答案,这些文本摘要是由多个知识库(无论是否结构化)中与问题相关的信息组合而成的。在Wikidata和DBpedia上的实验表明,我们的方法在域和KG无关的情况下,在单事实问题上取得了与其他方法相当的性能。结果为未来的工作提出了重要的问题,即如何从知识图中创建文本内容以实现答案提取。
{"title":"MuHeQA: Zero-shot question answering over multiple and heterogeneous knowledge bases","authors":"Carlos Badenes-Olmedo, Óscar Corcho","doi":"10.3233/sw-233379","DOIUrl":"https://doi.org/10.3233/sw-233379","url":null,"abstract":"There are two main limitations in most of the existing Knowledge Graph Question Answering (KGQA) algorithms. First, the approaches depend heavily on the structure and cannot be easily adapted to other KGs. Second, the availability and amount of additional domain-specific data in structured or unstructured formats has also proven to be critical in many of these systems. Such dependencies limit the applicability of KGQA systems and make their adoption difficult. A novel algorithm is proposed, MuHeQA, that alleviates both limitations by retrieving the answer from textual content automatically generated from KGs instead of queries over them. This new approach (1) works on one or several KGs simultaneously, (2) does not require training data what makes it is domain-independent, (3) enables the combination of knowledge graphs with unstructured information sources to build the answer, and (4) reduces the dependency on the underlying schema since it does not navigate through structured content but only reads property values. MuHeQA extracts answers from textual summaries created by combining information related to the question from multiple knowledge bases, be them structured or not. Experiments over Wikidata and DBpedia show that our approach achieves comparable performance to other approaches in single-fact questions while being domain and KG independent. Results raise important questions for future work about how the textual content that can be created from knowledge graphs enables answer extraction.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"86 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2023-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77148796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
LegalNERo: A linked corpus for named entity recognition in the Romanian legal domain LegalNERo:一个链接语料库,用于罗马尼亚法律领域的命名实体识别
3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-06-05 DOI: 10.3233/sw-233351
Vasile Păis,, Maria Mitrofan, Carol Luca Gasan, Alexandru Ianov, Corvin Ghit,ă, Vlad Silviu Coneschi, Andrei Onut,
LegalNERo is a manually annotated corpus for named entity recognition in the Romanian legal domain. It provides gold annotations for organizations, locations, persons, time expressions and legal resources mentioned in legal documents. Furthermore, GeoNames identifiers are provided. The resource is available in multiple formats, including span-based, token-based and RDF. The Linked Open Data version is available for both download and querying using SPARQL.
LegalNERo是一个手动注释的语料库,用于罗马尼亚法律领域的命名实体识别。为法律文件中提及的组织、地点、人物、时间表述、法律资源等提供黄金标注。此外,还提供了GeoNames标识符。该资源有多种格式,包括基于跨段的、基于令牌的和RDF。链接开放数据版本可以下载,也可以使用SPARQL进行查询。
{"title":"LegalNERo: A linked corpus for named entity recognition in the Romanian legal domain","authors":"Vasile Păis,, Maria Mitrofan, Carol Luca Gasan, Alexandru Ianov, Corvin Ghit,ă, Vlad Silviu Coneschi, Andrei Onut,","doi":"10.3233/sw-233351","DOIUrl":"https://doi.org/10.3233/sw-233351","url":null,"abstract":"LegalNERo is a manually annotated corpus for named entity recognition in the Romanian legal domain. It provides gold annotations for organizations, locations, persons, time expressions and legal resources mentioned in legal documents. Furthermore, GeoNames identifiers are provided. The resource is available in multiple formats, including span-based, token-based and RDF. The Linked Open Data version is available for both download and querying using SPARQL.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135702871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reason-able embeddings: Learning concept embeddings with a transferable neural reasoner 合理嵌入:使用可转移神经推理器学习概念嵌入
IF 3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-06-02 DOI: 10.3233/sw-233355
Dariusz Max Adamski, Jedrzej Potoniec
We present a novel approach for learning embeddings of ALC knowledge base concepts. The embeddings reflect the semantics of the concepts in such a way that it is possible to compute an embedding of a complex concept from the embeddings of its parts by using appropriate neural constructors. Embeddings for different knowledge bases are vectors in a shared vector space, shaped in such a way that approximate subsumption checking for arbitrarily complex concepts can be done by the same neural network, called a reasoner head, for all the knowledge bases. To underline this unique property of enabling reasoning directly on embeddings, we call them reason-able embeddings. We report the results of experimental evaluation showing that the difference in reasoning performance between training a separate reasoner head for each ontology and using a shared reasoner head, is negligible.
我们提出了一种学习ALC知识库概念嵌入的新方法。嵌入反映了概念的语义,通过使用适当的神经构造函数,可以从其部分的嵌入中计算出复杂概念的嵌入。不同知识库的嵌入是共享向量空间中的向量,其形状可以通过称为推理头的相同神经网络对所有知识库进行任意复杂概念的近似包容检查。为了强调直接对嵌入进行推理的独特特性,我们称之为合理嵌入。我们报告的实验评估结果表明,为每个本体训练单独的推理头和使用共享推理头之间的推理性能差异可以忽略不计。
{"title":"Reason-able embeddings: Learning concept embeddings with a transferable neural reasoner","authors":"Dariusz Max Adamski, Jedrzej Potoniec","doi":"10.3233/sw-233355","DOIUrl":"https://doi.org/10.3233/sw-233355","url":null,"abstract":"We present a novel approach for learning embeddings of ALC knowledge base concepts. The embeddings reflect the semantics of the concepts in such a way that it is possible to compute an embedding of a complex concept from the embeddings of its parts by using appropriate neural constructors. Embeddings for different knowledge bases are vectors in a shared vector space, shaped in such a way that approximate subsumption checking for arbitrarily complex concepts can be done by the same neural network, called a reasoner head, for all the knowledge bases. To underline this unique property of enabling reasoning directly on embeddings, we call them reason-able embeddings. We report the results of experimental evaluation showing that the difference in reasoning performance between training a separate reasoner head for each ontology and using a shared reasoner head, is negligible.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"23 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2023-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80504581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Neural axiom network for knowledge graph reasoning 知识图推理的神经公理网络
3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-05-29 DOI: 10.3233/sw-233276
Juan Li, Xiangnan Chen, Hongtao Yu, Jiaoyan Chen, Wen Zhang
Knowledge graph reasoning (KGR) aims to infer new knowledge or detect noises, which is essential for improving the quality of knowledge graphs. Recently, various KGR techniques, such as symbolic- and embedding-based methods, have been proposed and shown strong reasoning ability. Symbolic-based reasoning methods infer missing triples according to predefined rules or ontologies. Although rules and axioms have proven effective, it is difficult to obtain them. Embedding-based reasoning methods represent entities and relations as vectors, and complete KGs via vector computation. However, they mainly rely on structural information and ignore implicit axiom information not predefined in KGs but can be reflected in data. That is, each correct triple is also a logically consistent triple and satisfies all axioms. In this paper, we propose a novel NeuRal Axiom Network (NeuRAN) framework that combines explicit structural and implicit axiom information without introducing additional ontologies. Specifically, the framework consists of a KG embedding module that preserves the semantics of triples and five axiom modules that encode five kinds of implicit axioms. These axioms correspond to five typical object property expression axioms defined in OWL2, including ObjectPropertyDomain, ObjectPropertyRange, DisjointObjectProperties, IrreflexiveObjectProperty and AsymmetricObjectProperty. The KG embedding module and axiom modules compute the scores that the triple conforms to the semantics and the corresponding axioms, respectively. Compared with KG embedding models and CKRL, our method achieves comparable performance on noise detection and triple classification and achieves significant performance on link prediction. Compared with TransE and TransH, our method improves the link prediction performance on the Hits@1 metric by 22.0% and 20.8% on WN18RR-10% dataset, respectively.
知识图推理(Knowledge graph reasoning, KGR)旨在推断新知识或检测噪声,这是提高知识图质量的关键。近年来,各种KGR技术,如基于符号和基于嵌入的方法,已经被提出并显示出强大的推理能力。基于符号的推理方法根据预定义的规则或本体推断缺失的三元组。虽然规则和公理已被证明是有效的,但很难获得它们。基于嵌入的推理方法将实体和关系表示为向量,并通过向量计算完成KGs。然而,它们主要依赖于结构信息,而忽略了KGs中没有预定义但可以在数据中反映的隐含公理信息。也就是说,每个正确的三元组也是逻辑一致的三元组,并且满足所有公理。在本文中,我们提出了一个新的神经公理网络(NeuRAN)框架,它结合了显式结构和隐式公理信息,而不引入额外的本体。具体来说,该框架由一个保留三元组语义的KG嵌入模块和五个编码五种隐式公理的公理模块组成。这些公理对应于OWL2中定义的五个典型的对象属性表达式公理,包括ObjectPropertyDomain、ObjectPropertyRange、DisjointObjectProperties、IrreflexiveObjectProperty和AsymmetricObjectProperty。KG嵌入模块和公理模块分别计算三元组符合语义和相应公理的分数。与KG嵌入模型和CKRL模型相比,我们的方法在噪声检测和三重分类方面取得了相当的性能,在链路预测方面取得了显著的性能。与TransE和TransH相比,我们的方法在WN18RR-10%数据集上对Hits@1指标的链路预测性能分别提高了22.0%和20.8%。
{"title":"Neural axiom network for knowledge graph reasoning","authors":"Juan Li, Xiangnan Chen, Hongtao Yu, Jiaoyan Chen, Wen Zhang","doi":"10.3233/sw-233276","DOIUrl":"https://doi.org/10.3233/sw-233276","url":null,"abstract":"Knowledge graph reasoning (KGR) aims to infer new knowledge or detect noises, which is essential for improving the quality of knowledge graphs. Recently, various KGR techniques, such as symbolic- and embedding-based methods, have been proposed and shown strong reasoning ability. Symbolic-based reasoning methods infer missing triples according to predefined rules or ontologies. Although rules and axioms have proven effective, it is difficult to obtain them. Embedding-based reasoning methods represent entities and relations as vectors, and complete KGs via vector computation. However, they mainly rely on structural information and ignore implicit axiom information not predefined in KGs but can be reflected in data. That is, each correct triple is also a logically consistent triple and satisfies all axioms. In this paper, we propose a novel NeuRal Axiom Network (NeuRAN) framework that combines explicit structural and implicit axiom information without introducing additional ontologies. Specifically, the framework consists of a KG embedding module that preserves the semantics of triples and five axiom modules that encode five kinds of implicit axioms. These axioms correspond to five typical object property expression axioms defined in OWL2, including ObjectPropertyDomain, ObjectPropertyRange, DisjointObjectProperties, IrreflexiveObjectProperty and AsymmetricObjectProperty. The KG embedding module and axiom modules compute the scores that the triple conforms to the semantics and the corresponding axioms, respectively. Compared with KG embedding models and CKRL, our method achieves comparable performance on noise detection and triple classification and achieves significant performance on link prediction. Compared with TransE and TransH, our method improves the link prediction performance on the Hits@1 metric by 22.0% and 20.8% on WN18RR-10% dataset, respectively.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135792310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reuse of the FoodOn ontology in a knowledge base of food composition data 在食品成分数据知识库中重用FoodOn本体
3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-05-29 DOI: 10.3233/sw-233207
Katherine Thornton, Kenneth Seals-Nutt, Mika Matsuzaki, Damion Dooley
We describe our work to integrate the FoodOn ontology with our knowledge base of food composition data, WikiFCD. WikiFCD is knowledge base of structured data related to food composition and food items. With a goal to reuse FoodOn identifiers for food items, we imported a subset of the FoodOn ontology into the WikiFCD knowledge base. We aligned the import via a shared use of NCBI taxon identifiers for the taxon names of the plants from which the food items are derived. Reusing FoodOn benefits WikiFCD by allowing us to leverage the food item groupings that FoodOn contains. This integration also has potential future benefits for the FoodOn community due to the fact that WikiFCD provides food composition data at the food item level, and that WikiFCD is mapped to Wikidata and contains a SPARQL endpoint that supports federated queries. Federated queries across WikiFCD and Wikidata allow us to ask questions about food items that benefit from the cross-domain information of Wikidata, greatly increasing the breadth of possible data combinations.
我们描述了将FoodOn本体与我们的食品成分数据知识库WikiFCD集成的工作。WikiFCD是与食品成分和食品项目相关的结构化数据知识库。为了重用食品的FoodOn标识符,我们将FoodOn本体的一个子集导入WikiFCD知识库。我们通过共享使用NCBI分类单元标识符来对齐进口,这些分类单元名称来源于食物的植物。重用FoodOn对WikiFCD有利,因为它允许我们利用FoodOn所包含的食物项分组。由于WikiFCD在食品项目级别提供食品成分数据,并且WikiFCD映射到Wikidata并包含一个支持联邦查询的SPARQL端点,因此这种集成在未来对FoodOn社区也有潜在的好处。跨WikiFCD和Wikidata的联合查询允许我们询问有关食品的问题,这些问题受益于Wikidata的跨域信息,极大地增加了可能的数据组合的广度。
{"title":"Reuse of the FoodOn ontology in a knowledge base of food composition data","authors":"Katherine Thornton, Kenneth Seals-Nutt, Mika Matsuzaki, Damion Dooley","doi":"10.3233/sw-233207","DOIUrl":"https://doi.org/10.3233/sw-233207","url":null,"abstract":"We describe our work to integrate the FoodOn ontology with our knowledge base of food composition data, WikiFCD. WikiFCD is knowledge base of structured data related to food composition and food items. With a goal to reuse FoodOn identifiers for food items, we imported a subset of the FoodOn ontology into the WikiFCD knowledge base. We aligned the import via a shared use of NCBI taxon identifiers for the taxon names of the plants from which the food items are derived. Reusing FoodOn benefits WikiFCD by allowing us to leverage the food item groupings that FoodOn contains. This integration also has potential future benefits for the FoodOn community due to the fact that WikiFCD provides food composition data at the food item level, and that WikiFCD is mapped to Wikidata and contains a SPARQL endpoint that supports federated queries. Federated queries across WikiFCD and Wikidata allow us to ask questions about food items that benefit from the cross-domain information of Wikidata, greatly increasing the breadth of possible data combinations.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135791825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An ontology for maintenance activities and its application to data quality 维护活动本体及其在数据质量中的应用
3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-05-29 DOI: 10.3233/sw-233299
Caitlin Woods, Matt Selway, Tyler Bikaun, Markus Stumptner, Melinda Hodkiewicz
Maintenance of assets is a multi-million dollar cost each year for asset intensive organisations in the defence, manufacturing, resource and infrastructure sectors. These costs are tracked though maintenance work order (MWO) records. MWO records contain structured data for dates, costs, and asset identification and unstructured text describing the work required, for example ‘replace leaking pump’. Our focus in this paper is on data quality for maintenance activity terms in MWO records (e.g. replace, repair, adjust and inspect). We present two contributions in this paper. First, we propose a reference ontology for maintenance activity terms. We use natural language processing to identify seven core maintenance activity terms and their synonyms from 800,000 MWOs. We provide elucidations for these seven terms. Second, we demonstrate use of the reference ontology in an application-level ontology using an industrial use case. The end-to-end NLP-ontology pipeline identifies data quality issues with 55% of the MWO records for a centrifugal pump over 8 years. For the 33% of records where a verb was not provided in the unstructured text, the ontology can infer a relevant activity class. The selection of the maintenance activity terms is informed by the ISO 14224 and ISO 15926-4 standards and conforms to ISO/IEC 21838-2 Basic Formal Ontology (BFO). The reference and application ontologies presented here provide an example for how industrial organisations can augment their maintenance work management processes with ontological workflows to improve data quality.
对于国防、制造业、资源和基础设施领域的资产密集型企业来说,每年的资产维护成本高达数百万美元。这些费用通过维护工作订单(MWO)记录进行跟踪。MWO记录包含日期、成本和资产标识的结构化数据,以及描述所需工作的非结构化文本,例如“更换泄漏泵”。本文的重点是MWO记录中维护活动术语(如更换、修理、调整和检查)的数据质量。我们在这篇论文中提出了两个贡献。首先,我们提出了维护活动术语的参考本体。我们使用自然语言处理从80万个mwo中识别出7个核心维护活动术语及其同义词。我们对这七个术语作了说明。其次,我们通过一个工业用例演示了在应用级本体中引用本体的使用。端到端nlp本体管道识别了离心泵8年来55%的MWO记录的数据质量问题。对于在非结构化文本中没有提供动词的33%的记录,本体可以推断出相关的活动类。维护活动术语的选择依据ISO 14224和ISO 15926-4标准,并符合ISO/IEC 21838-2基本形式本体(BFO)。本文介绍的参考本体论和应用程序本体论为工业组织如何使用本体论工作流增强其维护工作管理流程以提高数据质量提供了一个示例。
{"title":"An ontology for maintenance activities and its application to data quality","authors":"Caitlin Woods, Matt Selway, Tyler Bikaun, Markus Stumptner, Melinda Hodkiewicz","doi":"10.3233/sw-233299","DOIUrl":"https://doi.org/10.3233/sw-233299","url":null,"abstract":"Maintenance of assets is a multi-million dollar cost each year for asset intensive organisations in the defence, manufacturing, resource and infrastructure sectors. These costs are tracked though maintenance work order (MWO) records. MWO records contain structured data for dates, costs, and asset identification and unstructured text describing the work required, for example ‘replace leaking pump’. Our focus in this paper is on data quality for maintenance activity terms in MWO records (e.g. replace, repair, adjust and inspect). We present two contributions in this paper. First, we propose a reference ontology for maintenance activity terms. We use natural language processing to identify seven core maintenance activity terms and their synonyms from 800,000 MWOs. We provide elucidations for these seven terms. Second, we demonstrate use of the reference ontology in an application-level ontology using an industrial use case. The end-to-end NLP-ontology pipeline identifies data quality issues with 55% of the MWO records for a centrifugal pump over 8 years. For the 33% of records where a verb was not provided in the unstructured text, the ontology can infer a relevant activity class. The selection of the maintenance activity terms is informed by the ISO 14224 and ISO 15926-4 standards and conforms to ISO/IEC 21838-2 Basic Formal Ontology (BFO). The reference and application ontologies presented here provide an example for how industrial organisations can augment their maintenance work management processes with ontological workflows to improve data quality.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135791829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data sharing in agricultural supply chains: Using semantics to enable sustainable food systems 农业供应链中的数据共享:利用语义实现可持续粮食系统
IF 3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-05-25 DOI: 10.3233/sw-233287
C. Brewster, N. Kalatzis, B. Nouwt, Han Kruiger, J. Verhoosel
The agrifood system faces a great many economic, social and environmental challenges. One of the biggest practical challenges has been to achieve greater data sharing throughout the agrifood system and the supply chain, both to inform other stakeholders about a product and equally to incentivise greater environmental sustainability. In this paper, a data sharing architecture is described built on three principles (a) reuse of existing semantic standards; (b) integration with legacy systems; and (c) a distributed architecture where stakeholders control access to their own data. The system has been developed based on the requirements of commercial users and is designed to allow queries across a federated network of agrifood stakeholders. The Ploutos semantic model is built on an integration of existing ontologies. The Ploutos architecture is built on a discovery directory and interoperability enablers, which use graph query patterns to traverse the network and collect the requisite data to be shared. The system is exemplified in the context of a pilot involving commercial stakeholders in the processed fruit sector. The data sharing approach is highly extensible with considerable potential for capturing sustainability related data.
农业粮食系统面临着许多经济、社会和环境方面的挑战。最大的实际挑战之一是在整个农业食品系统和供应链中实现更大的数据共享,既要告知其他利益相关者有关产品的信息,也要激励更大的环境可持续性。本文描述了基于三个原则的数据共享体系结构:(a)重用现有的语义标准;(b)与遗留系统结合;(c)分布式架构,利益相关者可以控制对自己数据的访问。该系统是根据商业用户的需求开发的,旨在允许跨农业食品利益相关者的联合网络进行查询。Ploutos语义模型建立在现有本体的集成上。Ploutos体系结构建立在发现目录和互操作性支持程序之上,这些支持程序使用图形查询模式遍历网络并收集要共享的必要数据。该系统在涉及加工水果部门商业利益攸关方的试点中得到例证。数据共享方法具有高度可扩展性,在获取与可持续性有关的数据方面具有相当大的潜力。
{"title":"Data sharing in agricultural supply chains: Using semantics to enable sustainable food systems","authors":"C. Brewster, N. Kalatzis, B. Nouwt, Han Kruiger, J. Verhoosel","doi":"10.3233/sw-233287","DOIUrl":"https://doi.org/10.3233/sw-233287","url":null,"abstract":"The agrifood system faces a great many economic, social and environmental challenges. One of the biggest practical challenges has been to achieve greater data sharing throughout the agrifood system and the supply chain, both to inform other stakeholders about a product and equally to incentivise greater environmental sustainability. In this paper, a data sharing architecture is described built on three principles (a) reuse of existing semantic standards; (b) integration with legacy systems; and (c) a distributed architecture where stakeholders control access to their own data. The system has been developed based on the requirements of commercial users and is designed to allow queries across a federated network of agrifood stakeholders. The Ploutos semantic model is built on an integration of existing ontologies. The Ploutos architecture is built on a discovery directory and interoperability enablers, which use graph query patterns to traverse the network and collect the requisite data to be shared. The system is exemplified in the context of a pilot involving commercial stakeholders in the processed fruit sector. The data sharing approach is highly extensible with considerable potential for capturing sustainability related data.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"1 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2023-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86453600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Engineering user-centered explanations to query answers in ontology-driven socio-technical systems 在本体驱动的社会技术系统中,以用户为中心的工程解释查询答案
IF 3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-05-22 DOI: 10.3233/sw-233297
J. C. Teze, José Paredes, Maria Vanina Martinez, Gerardo I. Simari
The role of explanations in intelligent systems has in the last few years entered the spotlight as AI-based solutions appear in an ever-growing set of applications. Though data-driven (or machine learning) techniques are often used as examples of how opaque (also called black box) approaches can lead to problems such as bias and general lack of explainability and interpretability, in reality these features are difficult to tame in general, even for approaches that are based on tools typically considered to be more amenable, like knowledge-based formalisms. In this paper, we continue a line of research and development towards building tools that facilitate the implementation of explainable and interpretable hybrid intelligent socio-technical systems, focusing on features that users can leverage to build explanations to their queries. In particular, we present the implementation of a recently-proposed application framework (and make available its source code) for developing such systems, and explore user-centered mechanisms for building explanations based both on the kinds of explanations required (such as counterfactual, contextual, etc.) and the inputs used for building them (coming from various sources, such as the knowledge base and lower-level data-driven modules). In order to validate our approach, we develop two use cases, one as a running example for detecting hate speech in social platforms and the other as an extension that also contemplates cyberbullying scenarios.
在过去的几年里,随着基于人工智能的解决方案出现在越来越多的应用程序中,解释在智能系统中的作用已经成为人们关注的焦点。虽然数据驱动(或机器学习)技术经常被用作不透明(也称为黑箱)方法如何导致偏见和普遍缺乏可解释性和可解释性等问题的例子,但实际上,这些特征通常很难被驯服,即使是基于通常被认为更容易接受的工具的方法,如基于知识的形式化。在本文中,我们继续研究和开发构建工具,以促进可解释和可解释混合智能社会技术系统的实现,重点关注用户可以利用的功能,以构建对其查询的解释。特别是,我们提出了一个最近提出的应用程序框架的实现(并提供其源代码),用于开发这样的系统,并探索以用户为中心的机制,基于所需的解释类型(如反事实、上下文等)和用于构建它们的输入(来自各种来源,如知识库和低级数据驱动模块)来构建解释。为了验证我们的方法,我们开发了两个用例,一个作为检测社交平台上的仇恨言论的运行示例,另一个作为考虑网络欺凌场景的扩展。
{"title":"Engineering user-centered explanations to query answers in ontology-driven socio-technical systems","authors":"J. C. Teze, José Paredes, Maria Vanina Martinez, Gerardo I. Simari","doi":"10.3233/sw-233297","DOIUrl":"https://doi.org/10.3233/sw-233297","url":null,"abstract":"The role of explanations in intelligent systems has in the last few years entered the spotlight as AI-based solutions appear in an ever-growing set of applications. Though data-driven (or machine learning) techniques are often used as examples of how opaque (also called black box) approaches can lead to problems such as bias and general lack of explainability and interpretability, in reality these features are difficult to tame in general, even for approaches that are based on tools typically considered to be more amenable, like knowledge-based formalisms. In this paper, we continue a line of research and development towards building tools that facilitate the implementation of explainable and interpretable hybrid intelligent socio-technical systems, focusing on features that users can leverage to build explanations to their queries. In particular, we present the implementation of a recently-proposed application framework (and make available its source code) for developing such systems, and explore user-centered mechanisms for building explanations based both on the kinds of explanations required (such as counterfactual, contextual, etc.) and the inputs used for building them (coming from various sources, such as the knowledge base and lower-level data-driven modules). In order to validate our approach, we develop two use cases, one as a running example for detecting hate speech in social platforms and the other as an extension that also contemplates cyberbullying scenarios.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"50 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2023-05-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88695048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Explanation Ontology: A general-purpose, semantic representation for supporting user-centered explanations 解释本体:用于支持以用户为中心的解释的通用语义表示
IF 3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-05-18 DOI: 10.3233/sw-233282
Shruthi Chari, O. Seneviratne, M. Ghalwash, Sola S. Shirai, Daniel Gruen, Pablo Meyer, P. Chakraborty, D. McGuinness
In the past decade, trustworthy Artificial Intelligence (AI) has emerged as a focus for the AI community to ensure better adoption of AI models, and explainable AI is a cornerstone in this area. Over the years, the focus has shifted from building transparent AI methods to making recommendations on how to make black-box or opaque machine learning models and their results more understandable by experts and non-expert users. In our previous work, to address the goal of supporting user-centered explanations that make model recommendations more explainable, we developed an Explanation Ontology (EO). The EO is a general-purpose representation that was designed to help system designers connect explanations to their underlying data and knowledge. This paper addresses the apparent need for improved interoperability to support a wider range of use cases. We expand the EO, mainly in the system attributes contributing to explanations, by introducing new classes and properties to support a broader range of state-of-the-art explainer models. We present the expanded ontology model, highlighting the classes and properties that are important to model a larger set of fifteen literature-backed explanation types that are supported within the expanded EO. We build on these explanation type descriptions to show how to utilize the EO model to represent explanations in five use cases spanning the domains of finance, food, and healthcare. We include competency questions that evaluate the EO’s capabilities to provide guidance for system designers on how to apply our ontology to their own use cases. This guidance includes allowing system designers to query the EO directly and providing them exemplar queries to explore content in the EO represented use cases. We have released this significantly expanded version of the Explanation Ontology at https://purl.org/heals/eo and updated our resource website, https://tetherless-world.github.io/explanation-ontology, with supporting documentation. Overall, through the EO model, we aim to help system designers be better informed about explanations and support these explanations that can be composed, given their systems’ outputs from various AI models, including a mix of machine learning, logical and explainer models, and different types of data and knowledge available to their systems.
在过去的十年里,可信赖的人工智能(AI)已经成为人工智能社区的焦点,以确保更好地采用人工智能模型,而可解释的人工智能是这一领域的基石。多年来,重点已经从构建透明的人工智能方法转移到就如何使黑箱或不透明的机器学习模型及其结果更容易被专家和非专业用户理解提出建议。在我们之前的工作中,为了解决支持以用户为中心的解释的目标,使模型建议更具可解释性,我们开发了一个解释本体(Explanation Ontology, EO)。EO是一种通用的表示形式,旨在帮助系统设计人员将解释与其基础数据和知识联系起来。本文提出了对改进互操作性的明显需求,以支持更广泛的用例。我们通过引入新的类和属性来支持更广泛的最先进的解释器模型,主要在有助于解释的系统属性中扩展EO。我们展示了扩展的本体模型,突出显示了类和属性,这些类和属性对于为扩展的EO中支持的15种文献支持的解释类型的更大集合建模很重要。我们以这些解释类型描述为基础,展示如何利用EO模型来表示跨越金融、食品和医疗保健领域的五个用例中的解释。我们包含了评估EO的能力的能力问题,为系统设计者提供关于如何将我们的本体应用到他们自己的用例中的指导。该指导包括允许系统设计人员直接查询EO,并为他们提供示例查询,以探索EO所表示用例中的内容。我们已经在https://purl.org/heals/eo上发布了这个重要的扩展版本的解释本体,并更新了我们的资源网站https://tetherless-world.github.io/explanation-ontology,并提供了支持文档。总的来说,通过EO模型,我们的目标是帮助系统设计师更好地了解解释,并支持这些可以组成的解释,因为他们的系统从各种人工智能模型输出,包括机器学习、逻辑和解释器模型的混合,以及系统可用的不同类型的数据和知识。
{"title":"Explanation Ontology: A general-purpose, semantic representation for supporting user-centered explanations","authors":"Shruthi Chari, O. Seneviratne, M. Ghalwash, Sola S. Shirai, Daniel Gruen, Pablo Meyer, P. Chakraborty, D. McGuinness","doi":"10.3233/sw-233282","DOIUrl":"https://doi.org/10.3233/sw-233282","url":null,"abstract":"In the past decade, trustworthy Artificial Intelligence (AI) has emerged as a focus for the AI community to ensure better adoption of AI models, and explainable AI is a cornerstone in this area. Over the years, the focus has shifted from building transparent AI methods to making recommendations on how to make black-box or opaque machine learning models and their results more understandable by experts and non-expert users. In our previous work, to address the goal of supporting user-centered explanations that make model recommendations more explainable, we developed an Explanation Ontology (EO). The EO is a general-purpose representation that was designed to help system designers connect explanations to their underlying data and knowledge. This paper addresses the apparent need for improved interoperability to support a wider range of use cases. We expand the EO, mainly in the system attributes contributing to explanations, by introducing new classes and properties to support a broader range of state-of-the-art explainer models. We present the expanded ontology model, highlighting the classes and properties that are important to model a larger set of fifteen literature-backed explanation types that are supported within the expanded EO. We build on these explanation type descriptions to show how to utilize the EO model to represent explanations in five use cases spanning the domains of finance, food, and healthcare. We include competency questions that evaluate the EO’s capabilities to provide guidance for system designers on how to apply our ontology to their own use cases. This guidance includes allowing system designers to query the EO directly and providing them exemplar queries to explore content in the EO represented use cases. We have released this significantly expanded version of the Explanation Ontology at https://purl.org/heals/eo and updated our resource website, https://tetherless-world.github.io/explanation-ontology, with supporting documentation. Overall, through the EO model, we aim to help system designers be better informed about explanations and support these explanations that can be composed, given their systems’ outputs from various AI models, including a mix of machine learning, logical and explainer models, and different types of data and knowledge available to their systems.","PeriodicalId":48694,"journal":{"name":"Semantic Web","volume":"295 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2023-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77091552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Semantic Web
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1