首页 > 最新文献

Data & Knowledge Engineering最新文献

英文 中文
Four decades of data & knowledge engineering: A bibliometric analysis and topic evolution study (1985–2024) 数据与知识工程的四十年:文献计量学分析和主题演变研究(1985-2024)
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-01 Epub Date: 2025-05-05 DOI: 10.1016/j.datak.2025.102462
Tatsawan Timakum , Soobin Lee , Dongha Kim , Min Song , Il-Yeol Song
The Data and Knowledge Engineering (DKE) journal has established a significant global research presence over four decades, substantially contributing to the advancement of data and knowledge engineering disciplines. This comprehensive bibliometric study analyzes the journal’s publications over the past 40 years (1985–2024), employing bibliographic records and citation data from Scopus, Web of Science (WoS), and ScienceDirect. By utilizing CiteSpace for citation and co-citation mapping and Dirichlet Multinomial Regression (DMR) topic modeling for trend analysis, the research provides a multifaceted examination of the journal’s scholarly landscape. Over its 40-year history, DKE has published 1951 articles, accumulating 53,594 citations. The study comprehensively explores key bibliometric dimensions, including influential authors, author networks, citation patterns, topic clusters, institutional contributions, and research funding sponsors, as well as evolution of topics, showing increasing, decreasing, or constant trends. Comprehensive analysis offers a meta-analytical perspective on DKE’s scholarly contributions, positioning the journal as a pioneering publication platform that advances critical knowledge and methodological innovations in data and knowledge engineering research domains. Through an in-depth examination of the journal’s publication trajectory, the study provides insights into the field’s scholarly evolution, highlighting DKE’s pivotal role in shaping academic discourse and technological understanding.
数据与知识工程(DKE)期刊在过去四十年中建立了重要的全球研究存在,为数据和知识工程学科的进步做出了重大贡献。这项全面的文献计量学研究分析了该期刊过去40年(1985-2024)的出版物,采用了来自Scopus、Web of Science (WoS)和ScienceDirect的书目记录和引文数据。本研究利用CiteSpace进行被引和共被引映射,利用Dirichlet多项式回归(DMR)主题建模进行趋势分析,对该期刊的学术格局进行了多方面的考察。建刊40年来,共发表文章1951篇,累计引用53594次。该研究全面探索了关键的文献计量维度,包括有影响力的作者、作者网络、引用模式、主题集群、机构贡献和研究资助赞助商,以及主题的演变,呈现出增加、减少或不变的趋势。综合分析为DKE的学术贡献提供了一个元分析的视角,将该期刊定位为一个开创性的出版平台,在数据和知识工程研究领域推进关键知识和方法创新。通过对期刊出版轨迹的深入研究,该研究提供了对该领域学术演变的见解,突出了DKE在塑造学术话语和技术理解方面的关键作用。
{"title":"Four decades of data & knowledge engineering: A bibliometric analysis and topic evolution study (1985–2024)","authors":"Tatsawan Timakum ,&nbsp;Soobin Lee ,&nbsp;Dongha Kim ,&nbsp;Min Song ,&nbsp;Il-Yeol Song","doi":"10.1016/j.datak.2025.102462","DOIUrl":"10.1016/j.datak.2025.102462","url":null,"abstract":"<div><div>The Data and Knowledge Engineering (DKE) journal has established a significant global research presence over four decades, substantially contributing to the advancement of data and knowledge engineering disciplines. This comprehensive bibliometric study analyzes the journal’s publications over the past 40 years (1985–2024), employing bibliographic records and citation data from Scopus, Web of Science (WoS), and ScienceDirect. By utilizing CiteSpace for citation and co-citation mapping and Dirichlet Multinomial Regression (DMR) topic modeling for trend analysis, the research provides a multifaceted examination of the journal’s scholarly landscape. Over its 40-year history, DKE has published 1951 articles, accumulating 53,594 citations. The study comprehensively explores key bibliometric dimensions, including influential authors, author networks, citation patterns, topic clusters, institutional contributions, and research funding sponsors, as well as evolution of topics, showing increasing, decreasing, or constant trends. Comprehensive analysis offers a meta-analytical perspective on DKE’s scholarly contributions, positioning the journal as a pioneering publication platform that advances critical knowledge and methodological innovations in data and knowledge engineering research domains. Through an in-depth examination of the journal’s publication trajectory, the study provides insights into the field’s scholarly evolution, highlighting DKE’s pivotal role in shaping academic discourse and technological understanding.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"159 ","pages":"Article 102462"},"PeriodicalIF":2.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144106307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ensemble model with combined feature set for Big data classification in IoT scenario 基于组合特征集的物联网场景大数据分类集成模型
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-01 Epub Date: 2025-04-17 DOI: 10.1016/j.datak.2025.102447
Harivardhagini S (Professor) , Pranavanand S (Associate Professor) , Raghuram A (Professor)
Sensor nodes that are wirelessly connected to the internet and several systems make up the Internet of Things system. Large volumes of data are often stored in big data, which complicates the classification process. There are many Big data classification strategies in use, but the main issues are the management of secure information as well as computational time. This paper's goal is to suggest a novel classification system for big data in Internet of Things networks that operates in four main phases. Particularly, the healthcare data is considered as the Big data perspective to solve the classification problem. Since the healthcare Big data is the revolutionary tool in this industry, it is becoming the most vital point of patient-centric care. Different data sources are aggregated in this Big data healthcare ecosystem. The first stage is data acquisition which takes place via Internet of Things through sensors. The second stage is improved DSig normalization for input data preprocessing. The third stage is MapReduce framework-based feature extraction for handling the Big data. This extract features like raw data, mutual information, information gain, and improved Renyi entropy. Finally, the fourth stage is an ensemble disease classification model by the combination of Recurrent Neural Network, Neural Network, and Improved Support Vector Machine for predicting normal and abnormal diseases. The suggested work is implemented by the Python tool, and the effectiveness, specificity, sensitivity, precision, and other factors of the results are assessed. The proposed ensemble model achieves superior precision of 0.9573 for the training rate of 90 % when compared to the traditional models.
无线连接到互联网和多个系统的传感器节点组成了物联网系统。大量的数据通常存储在大数据中,这使得分类过程变得复杂。目前有许多大数据分类策略在使用中,但主要问题是安全信息的管理以及计算时间。本文的目标是为物联网网络中的大数据提出一种新的分类系统,该系统分为四个主要阶段。特别是将医疗数据作为大数据视角来解决分类问题。由于医疗大数据是这个行业的革命性工具,它正在成为以患者为中心的医疗的最重要的一点。不同的数据源聚集在这个大数据医疗生态系统中。第一阶段是通过传感器通过物联网进行数据采集。第二阶段是改进的DSig规范化输入数据预处理。第三阶段是基于MapReduce框架的大数据特征提取。该方法提取了原始数据、互信息、信息增益和改进的人义熵等特征。最后,第四阶段是将递归神经网络、神经网络和改进的支持向量机相结合的疾病集成分类模型,用于预测正常和异常疾病。建议的工作由Python工具实现,并评估结果的有效性、特异性、灵敏度、精度和其他因素。与传统模型相比,该集成模型的训练精度达到0.9573,训练率达到90%。
{"title":"Ensemble model with combined feature set for Big data classification in IoT scenario","authors":"Harivardhagini S (Professor) ,&nbsp;Pranavanand S (Associate Professor) ,&nbsp;Raghuram A (Professor)","doi":"10.1016/j.datak.2025.102447","DOIUrl":"10.1016/j.datak.2025.102447","url":null,"abstract":"<div><div>Sensor nodes that are wirelessly connected to the internet and several systems make up the Internet of Things system. Large volumes of data are often stored in big data, which complicates the classification process. There are many Big data classification strategies in use, but the main issues are the management of secure information as well as computational time. This paper's goal is to suggest a novel classification system for big data in Internet of Things networks that operates in four main phases. Particularly, the healthcare data is considered as the Big data perspective to solve the classification problem. Since the healthcare Big data is the revolutionary tool in this industry, it is becoming the most vital point of patient-centric care. Different data sources are aggregated in this Big data healthcare ecosystem. The first stage is data acquisition which takes place via Internet of Things through sensors. The second stage is improved DSig normalization for input data preprocessing. The third stage is MapReduce framework-based feature extraction for handling the Big data. This extract features like raw data, mutual information, information gain, and improved Renyi entropy. Finally, the fourth stage is an ensemble disease classification model by the combination of Recurrent Neural Network, Neural Network, and Improved Support Vector Machine for predicting normal and abnormal diseases. The suggested work is implemented by the Python tool, and the effectiveness, specificity, sensitivity, precision, and other factors of the results are assessed. The proposed ensemble model achieves superior precision of 0.9573 for the training rate of 90 % when compared to the traditional models.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"159 ","pages":"Article 102447"},"PeriodicalIF":2.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144084758","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
G2MBCF: Enhanced Named Entity Recognition for sensitive entities identification G2MBCF:用于敏感实体识别的增强命名实体识别
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-01 Epub Date: 2025-04-26 DOI: 10.1016/j.datak.2025.102444
Weibin Tian , Kaiming Gu , Shihui Xiao , Junbo Zhang , Wei Cui
With the increasing growth of data, work on data security is becoming increasingly important. As the core of important data detection, the sensitive entities identification (SEI) problem has become a hot topic in natural language processing (NLP) science. Named Entity Recognition (NER) is the foundation of SEI, however, current studies treat SEI only as a special case of the NER problem. It lacks more detailed considerations of implicit links between entities and relations. In this paper, we propose a novel enhanced method called G2MBCF based on latent factor model (LFM). We use knowledge graph to represent the NER primary result with semantic structure. Then we use G2MBCF to inscribe entities and relations through a ER matrix to mine implicit connections. Experiments show that compared to existing NER methods, our method enhances Recall and Precision of SEI. We also studied the influence of parameters in the experiments.
随着数据的不断增长,数据安全工作变得越来越重要。敏感实体识别(SEI)问题作为重要数据检测的核心,已成为自然语言处理(NLP)科学研究的热点。命名实体识别(NER)是命名实体识别的基础,但目前的研究仅将命名实体识别作为命名实体识别问题的一个特例。它缺乏对实体和关系之间隐含联系的更详细的考虑。本文提出了一种基于潜在因子模型(LFM)的新型增强方法G2MBCF。我们使用知识图来表示具有语义结构的NER初级结果。然后,我们使用G2MBCF通过E - R矩阵来嵌入实体和关系,以挖掘隐式连接。实验表明,与现有的NER方法相比,我们的方法提高了SEI的查全率和查准率。我们还研究了实验中参数的影响。
{"title":"G2MBCF: Enhanced Named Entity Recognition for sensitive entities identification","authors":"Weibin Tian ,&nbsp;Kaiming Gu ,&nbsp;Shihui Xiao ,&nbsp;Junbo Zhang ,&nbsp;Wei Cui","doi":"10.1016/j.datak.2025.102444","DOIUrl":"10.1016/j.datak.2025.102444","url":null,"abstract":"<div><div>With the increasing growth of data, work on data security is becoming increasingly important. As the core of important data detection, the sensitive entities identification (SEI) problem has become a hot topic in natural language processing (NLP) science. Named Entity Recognition (NER) is the foundation of SEI, however, current studies treat SEI only as a special case of the NER problem. It lacks more detailed considerations of implicit links between entities and relations. In this paper, we propose a novel enhanced method called G2MBCF based on latent factor model (LFM). We use knowledge graph to represent the NER primary result with semantic structure. Then we use G2MBCF to inscribe entities and relations through a <span><math><mrow><mi>E</mi><mo>−</mo><mi>R</mi></mrow></math></span> matrix to mine implicit connections. Experiments show that compared to existing NER methods, our method enhances <span><math><mrow><mi>R</mi><mi>e</mi><mi>c</mi><mi>a</mi><mi>l</mi><mi>l</mi></mrow></math></span> and <span><math><mrow><mi>P</mi><mi>r</mi><mi>e</mi><mi>c</mi><mi>i</mi><mi>s</mi><mi>i</mi><mi>o</mi><mi>n</mi></mrow></math></span> of SEI. We also studied the influence of parameters in the experiments.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"159 ","pages":"Article 102444"},"PeriodicalIF":2.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143900094","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Customized long short-term memory architecture for multi-document summarization with improved text feature set 用于多文档摘要的定制化长短时记忆架构,具有改进的文本特征集
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-01 Epub Date: 2025-03-25 DOI: 10.1016/j.datak.2025.102440
Satya Deo , Debajyoty Banik , Prasant Kumar Pattnaik
One among the most crucial concerns in the domain of Natural Language Processing (NLP) is the Multi-Document Summarization (MDS) and in recent decades, the focus on this issue has risen massively. Hence, it is vital for the NLP community to provide effective and reliable MDS methods. Current deep learning-dependent MDS techniques rely on the extraordinary capacity of neural networks, in order to extract distinctive features. Motivated by this fact, we introduce a novel MDS technique, named as Customized Long Short-Term Memory-based Multi-Document Summarization using IBi-GRU (CLSTM-MDS+IBi-GRU), which includes the following working processes. Firstly, the input data gets converted into tokens by the Bi-directional Transformer (BERT) tokenizer. The features, such as Term Frequency- Inverse Document Frequency (TF-IDF), Bag of Words (BoW), thematic features and an improved aspect term-based feature are then extracted afterwards. Finally, the summarization process takes place by utilizing the concatenation of Customized Long Short-Term Memory (CLSTM) with a pre-eminent layer. Accurate and high-quality summary is provided via introducing this layer in the LSTM module and the Bi-GRU-based Inception module (IBi-GRU), which can capture long range dependences through parallel convolution. The outcomes of this work prove the superiority of our CLSTM-MDS in the Multi-Document Summarization task.
自然语言处理(NLP)领域中最重要的问题之一是多文档摘要(MDS),近几十年来,这一问题得到了广泛关注。因此,提供有效可靠的MDS方法对NLP社区至关重要。当前依赖深度学习的MDS技术依赖于神经网络的非凡能力,以提取显著特征。基于此,我们提出了一种新的基于IBi-GRU的基于定制长短期记忆的多文档摘要技术(CLSTM-MDS+IBi-GRU),包括以下工作流程。首先,输入数据通过双向转换器(BERT)标记器转换为标记。然后提取词频-逆文档频率(TF-IDF)、词包(BoW)、主题特征和改进的基于词的方面特征。最后,总结过程通过利用自定义长短期记忆(CLSTM)与卓越层的连接进行。通过在LSTM模块和基于bi - gru的Inception模块(IBi-GRU)中引入该层,可以提供准确和高质量的摘要,该模块可以通过并行卷积捕获远程依赖关系。研究结果证明了我们的CLSTM-MDS在多文档摘要任务中的优越性。
{"title":"Customized long short-term memory architecture for multi-document summarization with improved text feature set","authors":"Satya Deo ,&nbsp;Debajyoty Banik ,&nbsp;Prasant Kumar Pattnaik","doi":"10.1016/j.datak.2025.102440","DOIUrl":"10.1016/j.datak.2025.102440","url":null,"abstract":"<div><div>One <strong>a</strong>mong the most crucial concerns in the domain of Natural Language Processing (NLP) is the Multi-Document Summarization (MDS) and in recent decades, the focus on this issue has risen massively. Hence, it is vital for the NLP community to provide effective and reliable MDS methods. Current deep learning-dependent MDS techniques rely on the extraordinary capacity of neural networks, in order to extract distinctive features. Motivated by this fact, we introduce a novel MDS technique, named as Customized Long Short-Term Memory-based Multi-Document Summarization using IBi-GRU <strong>(</strong>CLSTM-MDS+IBi-GRU), which includes the following working processes. Firstly, the input data gets converted into tokens by the Bi-directional Transformer (BERT) tokenizer. The features, such as Term Frequency- Inverse Document Frequency (TF-IDF), Bag of Words (BoW), thematic features and an improved aspect term-based feature are then extracted afterwards. Finally, the summarization process takes place by utilizing the concatenation of Customized Long Short-Term Memory (CLSTM) with a pre-eminent layer. Accurate and high-quality summary is provided via introducing this layer in the LSTM module and the Bi-GRU-based Inception module (IBi-GRU), which can capture long range dependences through parallel convolution. The outcomes of this work prove the superiority of our CLSTM-MDS in the Multi-Document Summarization task.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"159 ","pages":"Article 102440"},"PeriodicalIF":2.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143800459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Philosophical reflections on conceptual modeling as communication 作为交流的概念建模的哲学思考
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-01 Epub Date: 2025-05-10 DOI: 10.1016/j.datak.2025.102453
Mattia Fumagalli , Giancarlo Guizzardi
Conceptual modeling is a complex and demanding task. It is a task centered around the challenge of representing a portion of the world in a way that is objective, understandable, shareable, and reusable by a community of practitioners, who rely on models to design and implement software or to clarify the concepts within a given domain. The difficulty of conceptual modeling stems from the inherent limitations of human representation abilities, which cannot fully capture the infinite richness and diversity of the world, nor the endless possibilities for description enabled by language. Significant effort has been invested in addressing these challenges, particularly in the creation of effective and reusable conceptual models, which have presented numerous difficulties. This paper explores conceptual modeling from a philosophical standpoint, proposing that conceptual models should not be viewed merely as the static representational output of an a priori activity, subject to modification only during a preliminary design phase. Instead, they should be seen as dynamic artifacts that require continuous design, adaptation, and evolution from their inception to their application, which may account for multiple purposes. The paper seeks to highlight the importance of understanding conceptual modeling primarily as an act of communication, rather than just a process of information transmission. It also aims to clarify the distinction between these two aspects and to examine the potential implications of adopting a communicative approach to modeling. These implications extend not only to the tools and methodologies used in modeling but also to the ethical considerations that arise from such an approach.
概念建模是一项复杂而艰巨的任务。这是一项围绕着以一种客观、可理解、可共享和可重用的方式表示世界的一部分的挑战的任务,这些实践者社区依赖模型来设计和实现软件或澄清给定领域内的概念。概念建模的困难源于人类表征能力的固有局限性,无法完全捕捉世界的无限丰富性和多样性,也无法通过语言实现描述的无限可能性。在解决这些挑战方面,特别是在创建有效和可重用的概念模型方面,已经投入了大量的努力,这带来了许多困难。本文从哲学的角度探讨了概念建模,提出概念模型不应仅仅被视为先验活动的静态表征输出,仅在初步设计阶段进行修改。相反,它们应该被视为动态的工件,需要从它们的开始到应用程序的持续设计、适应和进化,这可能有多种用途。本文试图强调将概念建模理解为一种沟通行为的重要性,而不仅仅是信息传递的过程。它还旨在澄清这两个方面之间的区别,并研究采用交际方法建模的潜在含义。这些含义不仅扩展到建模中使用的工具和方法,而且还扩展到从这种方法中产生的道德考虑。
{"title":"Philosophical reflections on conceptual modeling as communication","authors":"Mattia Fumagalli ,&nbsp;Giancarlo Guizzardi","doi":"10.1016/j.datak.2025.102453","DOIUrl":"10.1016/j.datak.2025.102453","url":null,"abstract":"<div><div>Conceptual modeling is a complex and demanding task. It is a task centered around the challenge of representing a portion of the world in a way that is objective, understandable, shareable, and reusable by a community of practitioners, who rely on models to design and implement software or to clarify the concepts within a given domain. The difficulty of conceptual modeling stems from the inherent limitations of human representation abilities, which cannot fully capture the infinite richness and diversity of the world, nor the endless possibilities for description enabled by language. Significant effort has been invested in addressing these challenges, particularly in the creation of effective and reusable conceptual models, which have presented numerous difficulties. This paper explores conceptual modeling from a philosophical standpoint, proposing that conceptual models should not be viewed merely as the static representational output of an a priori activity, subject to modification only during a preliminary design phase. Instead, they should be seen as dynamic artifacts that require continuous design, adaptation, and evolution from their inception to their application, which may account for multiple purposes. The paper seeks to highlight the importance of understanding conceptual modeling primarily as an act of communication, rather than just a process of information transmission. It also aims to clarify the distinction between these two aspects and to examine the potential implications of adopting a <em>communicative approach to modeling</em>. These implications extend not only to the tools and methodologies used in modeling but also to the ethical considerations that arise from such an approach.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"159 ","pages":"Article 102453"},"PeriodicalIF":2.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143942790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collaboration with GenAI in engineering research design 与GenAI合作进行工程研究设计
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-01 Epub Date: 2025-04-10 DOI: 10.1016/j.datak.2025.102445
Fazel Naghdy
Over the past five years, the fast development and use of generative artificial intelligence (GenAI) and large language models (LLMs) has ushered in a new era of study, teaching, and learning in many domains. The role that GenAIs can play in engineering research is addressed. The related previous works report on the potential of GenAIs in the literature review process. However, such potential is not demonstrated by case studies and practical examples. The previous works also do not address how GenAIs can assist with all the steps traditionally taken to design research. This study examines the effectiveness of collaboration with GenAIs at various stages of research design. It explores whether collaboration with GenAIs can result in more focused and comprehensive outcomes. A generalised approach for collaboration with AI tools in research design is proposed. A case study to develop a research design on the concept of “shared machine-human driving” is deployed to show the validity of the articulated concepts. The case study demonstrates both the pros and cons of collaboration with GenAIs. The results generated at each stage are rigorously validated and thoroughly examined to ensure they remain free from inaccuracies or hallucinations and align with the original research objectives. When necessary, the results are manually adjusted and refined to uphold their integrity and accuracy. The findings produced by the various GenAI models utilized in this study highlight the key attributes of generative artificial intelligence, namely speed, efficiency, and scope. However, they also underscore the critical importance of researcher oversight, as unexamined inferences and interpretations can render the results irrelevant or meaningless.
在过去的五年中,生成式人工智能(GenAI)和大型语言模型(llm)的快速发展和使用,在许多领域开创了一个研究、教学和学习的新时代。讨论了GenAIs在工程研究中可以发挥的作用。在文献综述的过程中对前人的相关工作进行了报道。然而,这种潜力并没有通过案例研究和实际例子来证明。以前的工作也没有解决GenAIs如何协助传统上采取的设计研究的所有步骤。本研究考察了在研究设计的各个阶段与GenAIs合作的有效性。它探讨了与GenAIs的合作是否能够产生更有针对性和更全面的结果。提出了一种在研究设计中与人工智能工具合作的通用方法。通过一个案例研究,对“人机共享驾驶”概念进行了研究设计,以展示所阐述概念的有效性。案例研究展示了与GenAIs合作的优点和缺点。每个阶段产生的结果都经过严格的验证和彻底的检查,以确保它们没有不准确或幻觉,并与最初的研究目标保持一致。必要时,人工调整和改进结果,以保持其完整性和准确性。本研究中使用的各种GenAI模型产生的结果突出了生成式人工智能的关键属性,即速度、效率和范围。然而,它们也强调了研究人员监督的重要性,因为未经检验的推论和解释可能使结果无关或毫无意义。
{"title":"Collaboration with GenAI in engineering research design","authors":"Fazel Naghdy","doi":"10.1016/j.datak.2025.102445","DOIUrl":"10.1016/j.datak.2025.102445","url":null,"abstract":"<div><div>Over the past five years, the fast development and use of generative artificial intelligence (GenAI) and large language models (LLMs) has ushered in a new era of study, teaching, and learning in many domains. The role that GenAIs can play in engineering research is addressed. The related previous works report on the potential of GenAIs in the literature review process. However, such potential is not demonstrated by case studies and practical examples. The previous works also do not address how GenAIs can assist with all the steps traditionally taken to design research. This study examines the effectiveness of collaboration with GenAIs at various stages of research design. It explores whether collaboration with GenAIs can result in more focused and comprehensive outcomes. A generalised approach for collaboration with AI tools in research design is proposed. A case study to develop a research design on the concept of “shared machine-human driving” is deployed to show the validity of the articulated concepts. The case study demonstrates both the pros and cons of collaboration with GenAIs. The results generated at each stage are rigorously validated and thoroughly examined to ensure they remain free from inaccuracies or hallucinations and align with the original research objectives. When necessary, the results are manually adjusted and refined to uphold their integrity and accuracy. The findings produced by the various GenAI models utilized in this study highlight the key attributes of generative artificial intelligence, namely speed, efficiency, and scope. However, they also underscore the critical importance of researcher oversight, as unexamined inferences and interpretations can render the results irrelevant or meaningless.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"159 ","pages":"Article 102445"},"PeriodicalIF":2.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143942791","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Conceptual design of multidimensional cubes with LLMs: An investigation 基于llm的多维数据集概念设计研究
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-01 Epub Date: 2025-05-07 DOI: 10.1016/j.datak.2025.102452
Stefano Rizzi, Matteo Francia, Enrico Gallinucci, Matteo Golfarelli
Large Language Models (LLMs) can simulate human linguistic capabilities, thus producing a disruptive impact across several domains, including software engineering. In this paper we focus on a specific scenario of software engineering, that of conceptual design of multidimensional data cubes. The goal is to evaluate the performance of LLMs (precisely, of ChatGPT-4o) in multidimensional conceptual design using the Dimensional Fact Model as a reference. To this end, we formulate nine research questions to (i) understand the competences of ChatGPT in multidimensional conceptual design, following either a supply- or a demand-driven approach, and (ii) investigate to what extent they can be improved via prompt engineering. After describing the research process in terms of base criteria, technological setting, input/output format, prompt templates, test cases, and metrics for evaluating the results, we discuss the output of the experiment. Our main conclusions are that (i) when prompts are enhanced with detailed procedural instructions and examples, the results produced significantly improve in all cases; and (ii) overall, ChatGPT is better at demand-driven design than at supply-driven design.
大型语言模型(llm)可以模拟人类的语言能力,从而在包括软件工程在内的多个领域产生破坏性影响。在本文中,我们关注软件工程的一个特定场景,即多维数据集的概念设计。目标是使用维度事实模型作为参考,评估llm(准确地说是chatgpt - 40)在多维概念设计中的性能。为此,我们制定了九个研究问题,以(i)了解ChatGPT在多维概念设计中的能力,遵循供应驱动或需求驱动的方法,以及(ii)调查通过快速工程可以在多大程度上改进它们。在从基本标准、技术设置、输入/输出格式、提示模板、测试用例和评估结果的度量等方面描述了研究过程之后,我们讨论了实验的输出。我们的主要结论是:(i)当用详细的程序说明和示例加强提示时,在所有情况下产生的结果都显着改善;(ii)总体而言,ChatGPT更擅长需求驱动的设计,而不是供应驱动的设计。
{"title":"Conceptual design of multidimensional cubes with LLMs: An investigation","authors":"Stefano Rizzi,&nbsp;Matteo Francia,&nbsp;Enrico Gallinucci,&nbsp;Matteo Golfarelli","doi":"10.1016/j.datak.2025.102452","DOIUrl":"10.1016/j.datak.2025.102452","url":null,"abstract":"<div><div>Large Language Models (LLMs) can simulate human linguistic capabilities, thus producing a disruptive impact across several domains, including software engineering. In this paper we focus on a specific scenario of software engineering, that of conceptual design of multidimensional data cubes. The goal is to evaluate the performance of LLMs (precisely, of ChatGPT-4o) in multidimensional conceptual design using the Dimensional Fact Model as a reference. To this end, we formulate nine research questions to (i) understand the competences of ChatGPT in multidimensional conceptual design, following either a supply- or a demand-driven approach, and (ii) investigate to what extent they can be improved via prompt engineering. After describing the research process in terms of base criteria, technological setting, input/output format, prompt templates, test cases, and metrics for evaluating the results, we discuss the output of the experiment. Our main conclusions are that (i) when prompts are enhanced with detailed procedural instructions and examples, the results produced significantly improve in all cases; and (ii) overall, ChatGPT is better at demand-driven design than at supply-driven design.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"159 ","pages":"Article 102452"},"PeriodicalIF":2.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143947575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Derived multi-objective function for latency sensitive-based cloud object storage system using hybrid heuristic algorithm 利用混合启发式算法推导了基于延迟敏感的云对象存储系统的多目标函数
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-01 Epub Date: 2025-04-10 DOI: 10.1016/j.datak.2025.102448
N Nataraj , RV Nataraj
Cloud Object Storage System (COSS) is capable of storing and retrieving a ton of unstructured data items called objects which act as a core cloud service for contemporary web-based applications. While sharing the data among different parties, privacy preservation becomes challenging. Research Problem: From day-to-day activities, a high volume of requests are served daily thus, it leads to cause the latency issues. In a cloud storage system, the adaption of a holistic approach helps the user to identify sensitive information and analyze the unwanted files/data. With evolving of Internet of Things (IoT) applications are latency-sensitive, which does not function well with these new ideas and platforms that are available today. Overall Purpose of the Study: Therefore, a novel latency-aware COSS is implemented with the aid of multi-objective functionalities to allocate and reallocate data efficiently in order to sustain the storage process in the cloud environment. Design of the Study: This goal is accomplished by implementing a hybrid meta-heuristic approach with the integration of the Mother Optimization Algorithm (MOA) with Dolphin Swarm Optimization (DSO) algorithm. The implemented hybrid optimization algorithm is called the Hybrid Dolphin Swarm-based Mother Optimization Algorithm (HDS-MOA). The HDS-MOA considers the objective function by considering constraints like throughput, latency, resource usage, and active servers during the data allocation process. While considering data reallocation process, the developed HDS-MOA algorithm is also performed by considering the multi-objective constraints like cost, makespan, and energy. The diverse experimental test is conducted to prove its effectiveness by comparing it with other existing methods for storing data efficiently across cloud networks. Major findings of results: In the configuration 3, the proposed HDS-MOA attains 31.11 %, 55.71 %, 55.71 %, and 68.21 % enhanced than the OSSperf, queuing theory, scheduling technique, and Monte Carlo-PSO based on the latency analysis. Overview of Interpretations and Conclusions: The developed HDS-MOA assured the better performance on the data is preserved in the optimal locations having appropriate access time and less latency that is highly essential for the cloud object storage. This supports to enhance the overall user experience by boosting the data retrieval. Limitations of this Study with Solutions: The ability of the proposed algorithm needs to enhance on balancing the multiple objectives such as performance, cost, and fault tolerance for optimally performing the operations in real-time that makes the system to be more efficient as well as responsive in the dynamic variations in the demand.
云对象存储系统(COSS)能够存储和检索大量被称为对象的非结构化数据项,这些数据项作为当代基于web的应用程序的核心云服务。在各方之间共享数据时,隐私保护变得具有挑战性。研究问题:从日常活动来看,每天都要处理大量的请求,因此会导致延迟问题。在云存储系统中,采用整体方法可以帮助用户识别敏感信息并分析不需要的文件/数据。随着物联网(IoT)的发展,应用程序对延迟敏感,这与今天可用的这些新想法和平台不能很好地配合。研究的总体目的:因此,为了维持云环境中的存储过程,在多目标功能的帮助下,实现了一种新的延迟感知的COSS,以有效地分配和重新分配数据。研究设计:该目标是通过实现一种混合元启发式方法来实现的,该方法将母体优化算法(MOA)与海豚群优化算法(DSO)相结合。所实现的混合优化算法被称为基于海豚群的混合母优化算法(HDS-MOA)。HDS-MOA通过在数据分配过程中考虑吞吐量、延迟、资源使用和活动服务器等约束来考虑目标函数。在考虑数据再分配过程的同时,开发的HDS-MOA算法还考虑了成本、完工时间和能量等多目标约束。通过将其与其他现有的跨云网络高效存储数据的方法进行比较,进行了多样化的实验测试,以证明其有效性。在配置3中,基于时延分析的HDS-MOA比OSSperf、排队论、调度技术和Monte Carlo-PSO分别提高了31.11%、55.71%、55.71%和68.21%。概述解释和结论:开发的HDS-MOA确保了数据保存在最佳位置的更好性能,具有适当的访问时间和更少的延迟,这对云对象存储至关重要。这有助于通过提高数据检索来增强整体用户体验。本研究与解决方案的局限性:本文提出的算法需要增强在性能、成本和容错等多个目标之间的平衡能力,以优化实时执行操作,使系统在需求动态变化中更加高效和响应。
{"title":"Derived multi-objective function for latency sensitive-based cloud object storage system using hybrid heuristic algorithm","authors":"N Nataraj ,&nbsp;RV Nataraj","doi":"10.1016/j.datak.2025.102448","DOIUrl":"10.1016/j.datak.2025.102448","url":null,"abstract":"<div><div>Cloud Object Storage System (COSS) is capable of storing and retrieving a ton of unstructured data items called objects which act as a core cloud service for contemporary web-based applications. While sharing the data among different parties, privacy preservation becomes challenging. <em>Research Problem:</em> From day-to-day activities, a high volume of requests are served daily thus, it leads to cause the latency issues. In a cloud storage system, the adaption of a holistic approach helps the user to identify sensitive information and analyze the unwanted files/data. With evolving of Internet of Things (IoT) applications are latency-sensitive, which does not function well with these new ideas and platforms that are available today. <em>Overall Purpose of the Study:</em> Therefore, a novel latency-aware COSS is implemented with the aid of multi-objective functionalities to allocate and reallocate data efficiently in order to sustain the storage process in the cloud environment. <em>Design of the Study:</em> This goal is accomplished by implementing a hybrid meta-heuristic approach with the integration of the Mother Optimization Algorithm (MOA) with Dolphin Swarm Optimization (DSO) algorithm. The implemented hybrid optimization algorithm is called the Hybrid Dolphin Swarm-based Mother Optimization Algorithm (HDS-MOA). The HDS-MOA considers the objective function by considering constraints like throughput, latency, resource usage, and active servers during the data allocation process. While considering data reallocation process, the developed HDS-MOA algorithm is also performed by considering the multi-objective constraints like cost, makespan, and energy. The diverse experimental test is conducted to prove its effectiveness by comparing it with other existing methods for storing data efficiently across cloud networks. <em>Major findings of results:</em> In the configuration 3, the proposed HDS-MOA attains 31.11 %, 55.71 %, 55.71 %, and 68.21 % enhanced than the OSSperf, queuing theory, scheduling technique, and Monte Carlo-PSO based on the latency analysis. <em>Overview of Interpretations and Conclusions:</em> The developed HDS-MOA assured the better performance on the data is preserved in the optimal locations having appropriate access time and less latency that is highly essential for the cloud object storage. This supports to enhance the overall user experience by boosting the data retrieval. <em>Limitations of this Study with Solutions:</em> The ability of the proposed algorithm needs to enhance on balancing the multiple objectives such as performance, cost, and fault tolerance for optimally performing the operations in real-time that makes the system to be more efficient as well as responsive in the dynamic variations in the demand.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"159 ","pages":"Article 102448"},"PeriodicalIF":2.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143859469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modelling process durations with gamma mixtures for right-censored data: Applications in customer clustering, pattern recognition, drift detection, and rationalisation 用伽马混合对右删节数据建模过程持续时间:在客户聚类、模式识别、漂移检测和合理化中的应用
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 Epub Date: 2025-03-14 DOI: 10.1016/j.datak.2025.102430
Lingkai Yang , Sally McClean , Kevin Burke , Mark Donnelly , Kashaf Khan
Customer modelling, particularly concerning length of stay or process duration, is vital for identifying customer patterns and optimising business processes. Recent advancements in computing and database technologies have revolutionised statistics and business process analytics by producing heterogeneous data that reflects diverse customer behaviours. Different models should be employed for distinct customer categories, culminating in an overall mixture model. Furthermore, some customers may remain “alive” at the conclusion of the observation period, meaning their journeys are incomplete, resulting in right-censored (RC) duration data. This combination of heterogeneous and right-censored data introduces complexity to process duration modelling and analysis. This paper presents a general approach to modelling process duration data using a gamma mixture model, where each gamma distribution represents a specific customer pattern. The model is adapted to account for RC data by modifying the likelihood function during model fitting. The paper explores three key application scenarios: (1) offline pattern clustering, which categorises customers who have completed their journeys; (2) online pattern tracking, which monitors and predicts customer behaviours in real-time; and (3) concept drift detection and rationalisation, which identifies shifts in customer patterns and explains their underlying causes. The proposed method has been validated using synthetically generated data and real-world data from a hospital billing process. In all instances, the fitted models effectively represented the data and demonstrated strong performance across the three application scenarios.
客户建模,特别是关于停留时间或流程持续时间的建模,对于识别客户模式和优化业务流程至关重要。计算和数据库技术的最新进展通过产生反映不同客户行为的异构数据,彻底改变了统计和业务流程分析。应该为不同的客户类别使用不同的模型,最终形成一个整体混合模型。此外,一些客户可能在观察期结束时仍然“活着”,这意味着他们的旅程是不完整的,从而导致持续时间数据的正确审查(RC)。这种异构和正确审查数据的组合给过程持续时间建模和分析带来了复杂性。本文介绍了使用gamma混合模型对过程持续时间数据建模的一般方法,其中每个gamma分布代表一个特定的客户模式。通过在模型拟合过程中修改似然函数,使模型适应于RC数据。本文探讨了三种关键应用场景:(1)线下模式聚类,对已完成旅程的客户进行分类;(2)在线模式跟踪,实时监控和预测客户行为;(3)概念漂移检测和合理化,识别客户模式的变化并解释其潜在原因。所提出的方法已经使用综合生成的数据和来自医院计费过程的真实数据进行了验证。在所有实例中,拟合模型都有效地表示了数据,并在三个应用场景中展示了强大的性能。
{"title":"Modelling process durations with gamma mixtures for right-censored data: Applications in customer clustering, pattern recognition, drift detection, and rationalisation","authors":"Lingkai Yang ,&nbsp;Sally McClean ,&nbsp;Kevin Burke ,&nbsp;Mark Donnelly ,&nbsp;Kashaf Khan","doi":"10.1016/j.datak.2025.102430","DOIUrl":"10.1016/j.datak.2025.102430","url":null,"abstract":"<div><div>Customer modelling, particularly concerning length of stay or process duration, is vital for identifying customer patterns and optimising business processes. Recent advancements in computing and database technologies have revolutionised statistics and business process analytics by producing heterogeneous data that reflects diverse customer behaviours. Different models should be employed for distinct customer categories, culminating in an overall mixture model. Furthermore, some customers may remain “alive” at the conclusion of the observation period, meaning their journeys are incomplete, resulting in right-censored (RC) duration data. This combination of heterogeneous and right-censored data introduces complexity to process duration modelling and analysis. This paper presents a general approach to modelling process duration data using a gamma mixture model, where each gamma distribution represents a specific customer pattern. The model is adapted to account for RC data by modifying the likelihood function during model fitting. The paper explores three key application scenarios: (1) offline pattern clustering, which categorises customers who have completed their journeys; (2) online pattern tracking, which monitors and predicts customer behaviours in real-time; and (3) concept drift detection and rationalisation, which identifies shifts in customer patterns and explains their underlying causes. The proposed method has been validated using synthetically generated data and real-world data from a hospital billing process. In all instances, the fitted models effectively represented the data and demonstrated strong performance across the three application scenarios.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"158 ","pages":"Article 102430"},"PeriodicalIF":2.7,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143654587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CriMOnto: A generalized domain-specific ontology for modeling procedural norms of the Lebanese criminal law CriMOnto:一个用于模拟黎巴嫩刑法程序规范的广义领域特定本体
IF 2.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 Epub Date: 2025-02-27 DOI: 10.1016/j.datak.2025.102419
Mirna El Ghosh , Hala Naja , Habib Abdulrab , Mohamad Khalil
Criminal (or penal) law regulates offenses, offenders, and legal punishments. Modeling criminal law is gaining much attention in the ontology engineering community. However, a significant aspect is neglected: the explicit representation of procedural knowledge. Procedural norms, such as regulative norms, are addressed to agents in the normative system. They govern the different interactions among these agents. In this study, we propose a formal and faithful representation of the procedural aspect of legal norms in the context of the Lebanese Criminal Code. A modular domain-specific ontology named CriMOnto is developed for this purpose. CriMOnto is grounded in the Unified Foundational Ontology (UFO) and the legal core ontology UFO-L by applying the Ontology-Driven Conceptual Modeling (ODCM) process. Conceptual Ontology Patterns (COPs) are reused from UFO and UFO-L to build the hierarchical and procedural content of the ontology. CriMOnto is validated as a formal ontology and evaluated using a dual evaluation approach. The potential use of CriMOnto for lightweight rule-based decision support is discussed in this study.
刑法规定了罪行、罪犯和法律惩罚。刑法建模在本体工程界受到广泛关注。然而,一个重要的方面被忽视了:程序知识的明确表示。程序性规范,如规范性规范,是针对规范体系中的行为主体的。它们控制着这些因子之间不同的相互作用。在本研究中,我们建议在黎巴嫩刑法的背景下正式和忠实地表示法律规范的程序方面。为此,开发了一个名为CriMOnto的模块化领域特定本体。CriMOnto以统一基础本体(UFO)和法律核心本体UFO- l为基础,应用本体驱动的概念建模(ODCM)过程。从UFO和UFO- l中重用概念本体模式(cop)来构建本体的层次化和程序化内容。CriMOnto作为形式化本体进行验证,并使用双重评估方法进行评估。本研究讨论了CriMOnto在轻量级基于规则的决策支持中的潜在用途。
{"title":"CriMOnto: A generalized domain-specific ontology for modeling procedural norms of the Lebanese criminal law","authors":"Mirna El Ghosh ,&nbsp;Hala Naja ,&nbsp;Habib Abdulrab ,&nbsp;Mohamad Khalil","doi":"10.1016/j.datak.2025.102419","DOIUrl":"10.1016/j.datak.2025.102419","url":null,"abstract":"<div><div>Criminal (or penal) law regulates offenses, offenders, and legal punishments. Modeling criminal law is gaining much attention in the ontology engineering community. However, a significant aspect is neglected: the explicit representation of procedural knowledge. Procedural norms, such as regulative norms, are addressed to agents in the normative system. They govern the different interactions among these agents. In this study, we propose a formal and faithful representation of the procedural aspect of legal norms in the context of the Lebanese Criminal Code. A modular domain-specific ontology named CriMOnto is developed for this purpose. CriMOnto is grounded in the Unified Foundational Ontology (UFO) and the legal core ontology UFO-L by applying the Ontology-Driven Conceptual Modeling (ODCM) process. Conceptual Ontology Patterns (COPs) are reused from UFO and UFO-L to build the hierarchical and procedural content of the ontology. CriMOnto is validated as a formal ontology and evaluated using a dual evaluation approach. The potential use of CriMOnto for lightweight rule-based decision support is discussed in this study.</div></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"158 ","pages":"Article 102419"},"PeriodicalIF":2.7,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143529128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Data & Knowledge Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1