首页 > 最新文献

Expert Systems最新文献

英文 中文
Comparison of nature-inspired algorithms in finite element-based metaheuristic optimisation of laminated shells 基于有限元的层状壳体元启发式优化中的自然启发算法比较
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-14 DOI: 10.1111/exsy.13620
Subham Pal, Kanak Kalita, Salil Haldar

This work presents a unique technique for optimising composite laminates used as structural components, which is critical for situations where failure might result in disastrous effects. Unlike traditional surrogate-based optimisation approaches, this methodology combines the accurate modelling capabilities of finite element (FE) analysis with the iterative refining capacity of metaheuristic algorithms. By combining these two methodologies, our method intends to improve the design process of laminated shell structures, assuring robustness and dependability is crucial. Compared to existing benchmark solutions, the current FE shows a <1% error for cylindrical and spherical shells. The prime objective of this study is to identify the optimum ply angles for attaining a high fundamental frequency. The problem is NP-hard because the possible ply angles span a wide range (±90°), making it difficult for optimisation algorithms to find a solution. Seven popular metaheuristic algorithms, namely the genetic algorithm (GA), the ant lion optimisation (ALO), the arithmetic optimisation algorithm (AOA), the dragonfly algorithm (DA), the grey wolf optimisation (GWO), the salp swarm optimisation (SSO), and the whale optimisation algorithm (WOA), are applied to and compared on a wide range of shell design problems. It assesses parameter sensitivity, discovering significant design elements that influence dynamic behaviour. Convergence studies demonstrate the superior performance of AOA, GWO, and WOA optimisers. Rigorous statistical comparisons assist practitioners in picking the best optimisation technique. FE-GWO, FE-DA, and FE-SSA methods surpass the other techniques as well as the layerwise optimisation strategy. The findings obtained, employing the GWO, DA, and SSA optimisers, demonstrate ~3% improvement over the existing literature. With respect to conventional layup designs (cross-ply and angle-ply), the current optimised designs are better by at least 0.43% and as much as 48.91%.

这项研究提出了一种独特的技术,用于优化用作结构部件的复合材料层压板,这对于失效可能导致灾难性后果的情况至关重要。与传统的代用优化方法不同,这种方法结合了有限元(FE)分析的精确建模能力和元启发式算法的迭代改进能力。通过结合这两种方法,我们的方法旨在改进层状壳体结构的设计过程,确保其稳健性和可靠性至关重要。与现有的基准解决方案相比,当前的 FE 对圆柱形和球形壳体的误差小于 1%。本研究的首要目标是确定获得高基频的最佳层叠角。这个问题具有 NP 难度,因为可能的层角跨度很大(±90°),使得优化算法很难找到解决方案。七种流行的元启发式算法,即遗传算法 (GA)、蚁狮优化算法 (ALO)、算术优化算法 (AOA)、蜻蜓算法 (DA)、灰狼优化算法 (GWO)、沙蜂群优化算法 (SSO) 和鲸鱼优化算法 (WOA),被应用于各种外壳设计问题并进行了比较。它评估了参数敏感性,发现了影响动态行为的重要设计元素。收敛性研究证明了 AOA、GWO 和 WOA 优化器的卓越性能。严格的统计比较有助于从业人员选择最佳优化技术。FE-GWO、FE-DA 和 FE-SSA 方法超越了其他技术以及分层优化策略。采用 GWO、DA 和 SSA 优化器得出的结果比现有文献提高了约 3%。与传统的层叠设计(交叉层叠和角层叠)相比,目前的优化设计至少提高了 0.43%,最高提高了 48.91%。
{"title":"Comparison of nature-inspired algorithms in finite element-based metaheuristic optimisation of laminated shells","authors":"Subham Pal,&nbsp;Kanak Kalita,&nbsp;Salil Haldar","doi":"10.1111/exsy.13620","DOIUrl":"10.1111/exsy.13620","url":null,"abstract":"<p>This work presents a unique technique for optimising composite laminates used as structural components, which is critical for situations where failure might result in disastrous effects. Unlike traditional surrogate-based optimisation approaches, this methodology combines the accurate modelling capabilities of finite element (FE) analysis with the iterative refining capacity of metaheuristic algorithms. By combining these two methodologies, our method intends to improve the design process of laminated shell structures, assuring robustness and dependability is crucial. Compared to existing benchmark solutions, the current FE shows a &lt;1% error for cylindrical and spherical shells. The prime objective of this study is to identify the optimum ply angles for attaining a high fundamental frequency. The problem is NP-hard because the possible ply angles span a wide range (±90°), making it difficult for optimisation algorithms to find a solution. Seven popular metaheuristic algorithms, namely the genetic algorithm (GA), the ant lion optimisation (ALO), the arithmetic optimisation algorithm (AOA), the dragonfly algorithm (DA), the grey wolf optimisation (GWO), the salp swarm optimisation (SSO), and the whale optimisation algorithm (WOA), are applied to and compared on a wide range of shell design problems. It assesses parameter sensitivity, discovering significant design elements that influence dynamic behaviour. Convergence studies demonstrate the superior performance of AOA, GWO, and WOA optimisers. Rigorous statistical comparisons assist practitioners in picking the best optimisation technique. FE-GWO, FE-DA, and FE-SSA methods surpass the other techniques as well as the layerwise optimisation strategy. The findings obtained, employing the GWO, DA, and SSA optimisers, demonstrate ~3% improvement over the existing literature. With respect to conventional layup designs (cross-ply and angle-ply), the current optimised designs are better by at least 0.43% and as much as 48.91%.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 9","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140981570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Blockchain-enabled decentralized service selection for QoS-aware cloud manufacturing 区块链支持的去中心化服务选择,用于服务质量感知云制造
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-09 DOI: 10.1111/exsy.13602
Ke Meng, Zhiyong Wu, Muhammad Bilal, Xiaoyu Xia, Xiaolong Xu

In recent years, cloud manufacturing has brought both opportunities and challenges to the manufacturing industry. Cloud manufacturing enables global manufacturing resources to be unified and shared, thus breaking down geographical constraints to enhance the level and efficiency of manufacturing. However, with the explosive growth of manufacturing resources and user demands, traditional cloud manufacturing platforms will face problems of insufficient computility, lack of real-time data and difficulties in securing user privacy during the service selection process. In this article, a blockchain-based decentralized cloud manufacturing service selection method is proposed, where the computility resource is deployed in multiple distributed nodes rather than the traditional centralized cloud manufacturing platform to solve the problem of insufficient computility. The credibility of the users is evaluated based on their performance on the contract and the PBFT consensus algorithm is improved based on the credibility of the users. In addition, a tri-chain blockchain data storage model is designed to ensure the security, real-time and transparency of data in the cloud manufacturing service selection process. The experimental results show that the method both speeds up service selection process and improves the quality of service selection results, and achieves a significant increase in manufacturing efficiency.

近年来,云制造给制造业带来了机遇和挑战。云制造实现了全球制造资源的统一和共享,打破了地域限制,提高了制造水平和效率。然而,随着制造资源和用户需求的爆发式增长,传统的云制造平台在服务选择过程中会面临计算能力不足、实时数据缺乏、用户隐私难以保障等问题。本文提出了一种基于区块链的去中心化云制造服务选择方法,将计算资源部署在多个分布式节点上,而非传统的中心化云制造平台,以解决计算能力不足的问题。根据用户在合约上的表现来评估用户的可信度,并根据用户的可信度改进 PBFT 共识算法。此外,还设计了三链区块链数据存储模型,以确保云制造服务选择过程中数据的安全性、实时性和透明性。实验结果表明,该方法既加快了服务选择过程,又提高了服务选择结果的质量,实现了制造效率的显著提升。
{"title":"Blockchain-enabled decentralized service selection for QoS-aware cloud manufacturing","authors":"Ke Meng,&nbsp;Zhiyong Wu,&nbsp;Muhammad Bilal,&nbsp;Xiaoyu Xia,&nbsp;Xiaolong Xu","doi":"10.1111/exsy.13602","DOIUrl":"10.1111/exsy.13602","url":null,"abstract":"<p>In recent years, cloud manufacturing has brought both opportunities and challenges to the manufacturing industry. Cloud manufacturing enables global manufacturing resources to be unified and shared, thus breaking down geographical constraints to enhance the level and efficiency of manufacturing. However, with the explosive growth of manufacturing resources and user demands, traditional cloud manufacturing platforms will face problems of insufficient computility, lack of real-time data and difficulties in securing user privacy during the service selection process. In this article, a blockchain-based decentralized cloud manufacturing service selection method is proposed, where the computility resource is deployed in multiple distributed nodes rather than the traditional centralized cloud manufacturing platform to solve the problem of insufficient computility. The credibility of the users is evaluated based on their performance on the contract and the PBFT consensus algorithm is improved based on the credibility of the users. In addition, a tri-chain blockchain data storage model is designed to ensure the security, real-time and transparency of data in the cloud manufacturing service selection process. The experimental results show that the method both speeds up service selection process and improves the quality of service selection results, and achieves a significant increase in manufacturing efficiency.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"42 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/exsy.13602","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140935262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lung cancer computed tomography image classification using Attention based Capsule Network with dispersed dynamic routing 利用基于注意力的胶囊网络和分散动态路由进行肺癌计算机断层扫描图像分类
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-08 DOI: 10.1111/exsy.13607
Ramya Paramasivam, Sujata N. Patil, Srinivas Konda, K. L. Hemalatha

Lung cancer is relying as one of the significant and leading cause for the deaths which are based on cancer. So, an effective diagnosis is a crucial step to save the patients who are all dying due to lung cancer. Moreover, the diagnosis must be performed based on the severity of lung cancer and the severity can be addressed with the help of an optimal classification approach. So, this research introduced an Attention based Capsule Network (A-Caps Net) with dispersed dynamic routing to perform in-depth classification of the disease affected partitions of the image and results in better classification results. The attention layer with dispersed dynamic routing evaluates the digit capsule from feature vector in a constant manner. As the first stage, data acquisitioned from datasets such as Lung Nodule Analysis-16 (LUNA-16), The Cancer Imaging Archive (TCIA) dataset and Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI). After acquisitioning data, pre-processing is done to enhance the resolution of the image using Generative Adversarial Network. The pre-processed output is given as output for extraction of features that takes place using GLCM and VGG-16 which extracts the low level features and high level features respectively. Finally, categorization of lung cancer is performed using Attention based Capsule Network (A-Caps Net) with dispersed dynamic routing which categorize the lung cancer as benign and malignant. The results obtained through experimental analysis exhibits that proposed approach attained better accuracy of 99.57%, 99.91% and 99.29% for LUNA-16, LIDC-IDRI and TCIA dataset respectively. The classification accuracy achieved by the proposed approach for LUNA-16 dataset is 99.57% which is comparably higher than DBN, 3D CNN, Squeeze Nodule Net and 3D-DCNN with multi-layered filter with accuracies of 99.16%, 97.17% and 94.1% respectively.

肺癌是导致癌症死亡的主要原因之一。因此,有效的诊断是挽救因肺癌而濒临死亡的患者的关键一步。此外,诊断必须根据肺癌的严重程度来进行,而严重程度可以通过最佳分类方法来解决。因此,本研究引入了一种基于注意力的胶囊网络(A-Capsle Network,A-Caps 网络),该网络具有分散的动态路由功能,可对图像中受疾病影响的部分进行深入分类,从而获得更好的分类结果。具有分散动态路由功能的注意力层以恒定的方式从特征向量中评估数字胶囊。第一阶段,从肺结节分析-16(LUNA-16)、癌症成像档案(TCIA)数据集和肺图像数据库联盟和图像数据库资源倡议(LIDC-IDRI)等数据集中获取数据。获取数据后,使用生成对抗网络进行预处理,以提高图像的分辨率。预处理后的输出将作为提取特征的输出,提取特征时使用 GLCM 和 VGG-16,分别提取低层次特征和高层次特征。最后,利用基于注意力的胶囊网络(A-Caps Net)和分散的动态路由对肺癌进行分类,将肺癌分为良性和恶性。实验分析结果表明,在 LUNA-16、LIDC-IDRI 和 TCIA 数据集上,所提出的方法分别达到了 99.57%、99.91% 和 99.29% 的较高准确率。拟议方法在 LUNA-16 数据集上的分类准确率为 99.57%,高于 DBN、3D CNN、挤压结节网和带有多层滤波器的 3D-DCNN 的准确率(分别为 99.16%、97.17% 和 94.1%)。
{"title":"Lung cancer computed tomography image classification using Attention based Capsule Network with dispersed dynamic routing","authors":"Ramya Paramasivam,&nbsp;Sujata N. Patil,&nbsp;Srinivas Konda,&nbsp;K. L. Hemalatha","doi":"10.1111/exsy.13607","DOIUrl":"10.1111/exsy.13607","url":null,"abstract":"<p>Lung cancer is relying as one of the significant and leading cause for the deaths which are based on cancer. So, an effective diagnosis is a crucial step to save the patients who are all dying due to lung cancer. Moreover, the diagnosis must be performed based on the severity of lung cancer and the severity can be addressed with the help of an optimal classification approach. So, this research introduced an Attention based Capsule Network (A-Caps Net) with dispersed dynamic routing to perform in-depth classification of the disease affected partitions of the image and results in better classification results. The attention layer with dispersed dynamic routing evaluates the digit capsule from feature vector in a constant manner. As the first stage, data acquisitioned from datasets such as Lung Nodule Analysis-16 (LUNA-16), The Cancer Imaging Archive (TCIA) dataset and Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI). After acquisitioning data, pre-processing is done to enhance the resolution of the image using Generative Adversarial Network. The pre-processed output is given as output for extraction of features that takes place using GLCM and VGG-16 which extracts the low level features and high level features respectively. Finally, categorization of lung cancer is performed using Attention based Capsule Network (A-Caps Net) with dispersed dynamic routing which categorize the lung cancer as benign and malignant. The results obtained through experimental analysis exhibits that proposed approach attained better accuracy of 99.57%, 99.91% and 99.29% for LUNA-16, LIDC-IDRI and TCIA dataset respectively. The classification accuracy achieved by the proposed approach for LUNA-16 dataset is 99.57% which is comparably higher than DBN, 3D CNN, Squeeze Nodule Net and 3D-DCNN with multi-layered filter with accuracies of 99.16%, 97.17% and 94.1% respectively.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 9","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140935286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Doc-KG: Unstructured documents to knowledge graph construction, identification and validation with Wikidata Doc-KG:利用维基数据从非结构化文档到知识图谱的构建、识别和验证
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-08 DOI: 10.1111/exsy.13617
Muhammad Salman, Armin Haller, Sergio J. Rodríguez Méndez, Usman Naseem

The exponential growth of textual data in the digital era underlines the pivotal role of Knowledge Graphs (KGs) in effectively storing, managing, and utilizing this vast reservoir of information. Despite the copious amounts of text available on the web, a significant portion remains unstructured, presenting a substantial barrier to the automatic construction and enrichment of KGs. To address this issue, we introduce an enhanced Doc-KG model, a sophisticated approach designed to transform unstructured documents into structured knowledge by generating local KGs and mapping these to a target KG, such as Wikidata. Our model innovatively leverages syntactic information to extract entities and predicates efficiently, integrating them into triples with improved accuracy. Furthermore, the Doc-KG model's performance surpasses existing methodologies by utilizing advanced algorithms for both the extraction of triples and their subsequent identification within Wikidata, employing Wikidata's Unified Resource Identifiers for precise mapping. This dual capability not only facilitates the construction of KGs directly from unstructured texts but also enhances the process of identifying triple mentions within Wikidata, marking a significant advancement in the domain. Our comprehensive evaluation, conducted using the renowned WebNLG benchmark dataset, reveals the Doc-KG model's superior performance in triple extraction tasks, achieving an unprecedented accuracy rate of 86.64%. In the domain of triple identification, the model demonstrated exceptional efficacy by mapping 61.35% of the local KG to Wikidata, thereby contributing 38.65% of novel information for KG enrichment. A qualitative analysis based on a manually annotated dataset further confirms the model's excellence, outshining baseline methods in extracting high-fidelity triples. This research embodies a novel contribution to the field of knowledge extraction and management, offering a robust framework for the semantic structuring of unstructured data and paving the way for the next generation of KGs.

数字时代文本数据的指数级增长凸显了知识图谱(KG)在有效存储、管理和利用这一庞大信息库方面的关键作用。尽管网络上有大量的文本,但其中很大一部分仍然是非结构化的,这对自动构建和丰富知识图谱构成了巨大的障碍。为了解决这个问题,我们引入了一个增强的 Doc-KG 模型,这是一种复杂的方法,旨在通过生成本地 KG 并将其映射到目标 KG(如 Wikidata),将非结构化文档转化为结构化知识。我们的模型创新性地利用语法信息来高效提取实体和谓词,并将它们整合到三元组中,从而提高了准确性。此外,Doc-KG 模型的性能超越了现有方法,它利用先进的算法提取三元组,并随后在 Wikidata 中进行识别,采用 Wikidata 的统一资源标识符进行精确映射。这种双重能力不仅有助于直接从非结构化文本中构建KG,还增强了维基数据中三元提及的识别过程,标志着该领域的重大进步。我们使用著名的 WebNLG 基准数据集进行了全面评估,结果显示 Doc-KG 模型在三重提取任务中表现出色,准确率达到前所未有的 86.64%。在三重识别领域,该模型表现出了卓越的功效,将 61.35% 的本地 KG 映射到了 Wikidata,从而为丰富 KG 提供了 38.65% 的新信息。基于人工注释数据集的定性分析进一步证实了该模型的卓越性,在提取高保真三元组方面超越了基线方法。这项研究体现了对知识提取和管理领域的新贡献,为非结构化数据的语义结构化提供了一个强大的框架,并为下一代知识库铺平了道路。
{"title":"Doc-KG: Unstructured documents to knowledge graph construction, identification and validation with Wikidata","authors":"Muhammad Salman,&nbsp;Armin Haller,&nbsp;Sergio J. Rodríguez Méndez,&nbsp;Usman Naseem","doi":"10.1111/exsy.13617","DOIUrl":"10.1111/exsy.13617","url":null,"abstract":"<p>The exponential growth of textual data in the digital era underlines the pivotal role of Knowledge Graphs (KGs) in effectively storing, managing, and utilizing this vast reservoir of information. Despite the copious amounts of text available on the web, a significant portion remains unstructured, presenting a substantial barrier to the automatic construction and enrichment of KGs. To address this issue, we introduce an enhanced Doc-KG model, a sophisticated approach designed to transform unstructured documents into structured knowledge by generating local KGs and mapping these to a target KG, such as Wikidata. Our model innovatively leverages syntactic information to extract entities and predicates efficiently, integrating them into triples with improved accuracy. Furthermore, the Doc-KG model's performance surpasses existing methodologies by utilizing advanced algorithms for both the extraction of triples and their subsequent identification within Wikidata, employing Wikidata's Unified Resource Identifiers for precise mapping. This dual capability not only facilitates the construction of KGs directly from unstructured texts but also enhances the process of identifying triple mentions within Wikidata, marking a significant advancement in the domain. Our comprehensive evaluation, conducted using the renowned WebNLG benchmark dataset, reveals the Doc-KG model's superior performance in triple extraction tasks, achieving an unprecedented accuracy rate of 86.64%. In the domain of triple identification, the model demonstrated exceptional efficacy by mapping 61.35% of the local KG to Wikidata, thereby contributing 38.65% of novel information for KG enrichment. A qualitative analysis based on a manually annotated dataset further confirms the model's excellence, outshining baseline methods in extracting high-fidelity triples. This research embodies a novel contribution to the field of knowledge extraction and management, offering a robust framework for the semantic structuring of unstructured data and paving the way for the next generation of KGs.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 9","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/exsy.13617","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140931734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enterprise violation risk deduction combining generative AI and event evolution graph 结合生成式人工智能和事件演化图的企业违规风险推断
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-08 DOI: 10.1111/exsy.13622
Chao Zhong, Pengjun Li, Jinlong Wang, Xiaoyun Xiong, Zhihan Lv, Xiaochen Zhou, Qixin Zhao

In the current realms of scientific research and commercial applications, the risk inference of regulatory violations by publicly listed enterprises has attracted considerable attention. However, there are some problems in the existing research on the deduction and prediction of violation risk of listed enterprises, such as the lack of analysis of the causal logic association between violation events, the low interpretability and effectiveness of the deduction and the lack of training data. To solve these problems, we propose a framework for enterprise violation risk deduction based on generative AI and event evolution graphs. First, the generative AI technology was used to generate a new text summary of the lengthy and complex enterprise violation announcement to realize a concise overview of the violation matters. Second, by fine-tuning the generative AI model, an event entity and causality extraction framework based on automated data augmentation are proposed, and the UIE (Unified Structure Generation for Universal Information Extraction) event entity extraction model is used to create the event entity extraction for listed enterprises ‘violations. Then, a causality extraction model CDDP-GAT (Event Causality Extraction Based on Chinese Dictionary and Dependency Parsing of GAT) is proposed. This model aims to identify and analyse the causal links between corporate breaches, thereby deepening the understanding of the event logic. Then, the merger of similar events was realized, and the causal correlation weights between enterprise violation-related events were evaluated. Finally, the listed enterprise's violation risk event evolution graph was constructed, and the enterprise violation risk deduction was carried out to form an expert system of financial violations. The deduction results show that the method can effectively reveal signs of enterprise violations and adverse consequences.

在当前的科学研究和商业应用领域,上市企业违规风险推断已引起了广泛关注。然而,现有的上市企业违规风险推演和预测研究存在一些问题,如缺乏对违规事件之间因果逻辑关联的分析、推演的可解释性和有效性较低、缺乏训练数据等。为解决这些问题,我们提出了基于生成式人工智能和事件演化图的企业违规风险推演框架。首先,利用生成式人工智能技术将冗长复杂的企业违规公告生成新的文本摘要,实现对违规事项的简明概述。其次,通过对生成式人工智能模型进行微调,提出了基于数据自动增量的事件实体和因果关系抽取框架,并利用UIE(Unified Structure Generation for Universal Information Extraction)事件实体抽取模型创建了上市企业'违规'事件实体抽取。然后,提出了一个因果关系提取模型 CDDP-GAT(基于中文词典和 GAT 依赖关系解析的事件因果关系提取)。该模型旨在识别和分析企业违规行为之间的因果联系,从而加深对事件逻辑的理解。然后,实现了相似事件的合并,并评估了企业违规相关事件之间的因果关联权重。最后,构建上市企业违规风险事件演化图,进行企业违规风险演绎,形成财务违规专家系统。演绎结果表明,该方法能有效揭示企业违规迹象及不良后果。
{"title":"Enterprise violation risk deduction combining generative AI and event evolution graph","authors":"Chao Zhong,&nbsp;Pengjun Li,&nbsp;Jinlong Wang,&nbsp;Xiaoyun Xiong,&nbsp;Zhihan Lv,&nbsp;Xiaochen Zhou,&nbsp;Qixin Zhao","doi":"10.1111/exsy.13622","DOIUrl":"10.1111/exsy.13622","url":null,"abstract":"<p>In the current realms of scientific research and commercial applications, the risk inference of regulatory violations by publicly listed enterprises has attracted considerable attention. However, there are some problems in the existing research on the deduction and prediction of violation risk of listed enterprises, such as the lack of analysis of the causal logic association between violation events, the low interpretability and effectiveness of the deduction and the lack of training data. To solve these problems, we propose a framework for enterprise violation risk deduction based on generative AI and event evolution graphs. First, the generative AI technology was used to generate a new text summary of the lengthy and complex enterprise violation announcement to realize a concise overview of the violation matters. Second, by fine-tuning the generative AI model, an event entity and causality extraction framework based on automated data augmentation are proposed, and the UIE (Unified Structure Generation for Universal Information Extraction) event entity extraction model is used to create the event entity extraction for listed enterprises ‘violations. Then, a causality extraction model CDDP-GAT (Event Causality Extraction Based on Chinese Dictionary and Dependency Parsing of GAT) is proposed. This model aims to identify and analyse the causal links between corporate breaches, thereby deepening the understanding of the event logic. Then, the merger of similar events was realized, and the causal correlation weights between enterprise violation-related events were evaluated. Finally, the listed enterprise's violation risk event evolution graph was constructed, and the enterprise violation risk deduction was carried out to form an expert system of financial violations. The deduction results show that the method can effectively reveal signs of enterprise violations and adverse consequences.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"42 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140931811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Nonlinear dynamical system approximation and adaptive control based on hybrid-feed-forward recurrent neural network: Simulation and stability analysis 基于混合前馈递归神经网络的非线性动力系统近似与自适应控制:仿真与稳定性分析
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-05 DOI: 10.1111/exsy.13619
R. Shobana, Rajesh Kumar, Bhavnesh Jaint

We proposed an online identification and adaptive control framework for the nonlinear dynamical systems using a novel hybrid-feed-forward recurrent neural network (HFRNN) model. The HFRNN is a combination of a feed-forward neural network (FFNN) and a local recurrent neural network (LRNN). We aim to leverage the simplicity of FFNN and the effectiveness of RNN to capture changing dynamics accurately and design an indirect adaptive control scheme. To derive the weights update equations, we have applied the gradient-descent-based Back-Propagation (BP) technique, and the stability of the proposed learning strategy is proven using the Lyapunov stability principles. We also compared the proposed method's results with those of the Jordan network-based controller (JNC) and the local recurrent network-based controller (LRNC) in the simulation examples. The results demonstrate that our approach performs satisfactorily, even in the presence of disturbance signals.

我们提出了一种使用新型混合前馈递归神经网络(HFRNN)模型的非线性动力系统在线识别和自适应控制框架。HFRNN 是前馈神经网络 (FFNN) 和局部递归神经网络 (LRNN) 的组合。我们旨在利用前馈神经网络的简单性和局部递归神经网络的有效性来准确捕捉动态变化,并设计一种间接自适应控制方案。为了推导权值更新方程,我们应用了基于梯度下降的反向传播(BP)技术,并利用 Lyapunov 稳定性原理证明了拟议学习策略的稳定性。我们还在仿真实例中比较了所提方法与基于约旦网络的控制器(JNC)和基于局部递归网络的控制器(LRNC)的结果。结果表明,即使存在干扰信号,我们的方法也能令人满意地运行。
{"title":"Nonlinear dynamical system approximation and adaptive control based on hybrid-feed-forward recurrent neural network: Simulation and stability analysis","authors":"R. Shobana,&nbsp;Rajesh Kumar,&nbsp;Bhavnesh Jaint","doi":"10.1111/exsy.13619","DOIUrl":"10.1111/exsy.13619","url":null,"abstract":"<p>We proposed an online identification and adaptive control framework for the nonlinear dynamical systems using a novel hybrid-feed-forward recurrent neural network (HFRNN) model. The HFRNN is a combination of a feed-forward neural network (FFNN) and a local recurrent neural network (LRNN). We aim to leverage the simplicity of FFNN and the effectiveness of RNN to capture changing dynamics accurately and design an indirect adaptive control scheme. To derive the weights update equations, we have applied the gradient-descent-based Back-Propagation (BP) technique, and the stability of the proposed learning strategy is proven using the Lyapunov stability principles. We also compared the proposed method's results with those of the Jordan network-based controller (JNC) and the local recurrent network-based controller (LRNC) in the simulation examples. The results demonstrate that our approach performs satisfactorily, even in the presence of disturbance signals.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 9","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-05-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140884928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning-based aggregate analysis to identify cut-off points for decision-making in pancreatic cancer detection 基于深度学习的总体分析确定胰腺癌检测决策的临界点
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-04-25 DOI: 10.1111/exsy.13614
Gintautas Dzemyda, Olga Kurasova, Viktor Medvedev, Aušra Šubonienė, Aistė Gulla, Artūras Samuilis, Džiugas Jagminas, Kęstutis Strupas

This study addresses the problem of detecting pancreatic cancer by classifying computed tomography (CT) images into cancerous and non-cancerous classes using the proposed deep learning-based aggregate analysis framework. The application of deep learning, as a branch of machine learning and artificial intelligence, to specific medical challenges can lead to the early detection of diseases, thus accelerating the process towards timely and effective intervention. The concept of classification is to reasonably select an optimal cut-off point, which is used as a threshold for evaluating the model results. The choice of this point is key to ensure efficient evaluation of the classification results, which directly affects the diagnostic accuracy. A significant aspect of this research is the incorporation of private CT images from Vilnius University Hospital Santaros Klinikos, combined with publicly available data sets. To investigate the capabilities of the deep learning-based framework and to maximize pancreatic cancer diagnostic performance, experimental studies were carried out combining data from different sources. Classification accuracy metrics such as the Youden index, (0, 1)-criterion, Matthew's correlation coefficient, the F1 score, LR+, LR−, balanced accuracy, and g-mean were used to find the optimal cut-off point in order to balance sensitivity and specificity. By carefully analyzing and comparing the obtained results, we aim to develop a reliable system that will not only improve the accuracy of pancreatic cancer detection but also have wider application in the early diagnosis of other malignancies.

本研究利用所提出的基于深度学习的聚合分析框架,将计算机断层扫描(CT)图像分为癌症和非癌症两类,从而解决了检测胰腺癌的问题。深度学习是机器学习和人工智能的一个分支,将其应用于特定的医疗挑战,可实现疾病的早期检测,从而加快及时有效的干预进程。分类的概念是合理选择一个最佳临界点,作为评估模型结果的阈值。该点的选择是确保高效评估分类结果的关键,它直接影响到诊断的准确性。这项研究的一个重要方面是将维尔纽斯大学 Santaros Klinikos 医院的私人 CT 图像与公开数据集相结合。为了研究基于深度学习的框架的能力,并最大限度地提高胰腺癌诊断性能,我们结合不同来源的数据进行了实验研究。实验中使用了尤登指数、(0,1)标准、马太相关系数、F1 分数、LR+、LR-、平衡准确度和 g-mean 等分类准确度指标,以找到平衡灵敏度和特异性的最佳临界点。通过仔细分析和比较所获得的结果,我们希望开发出一种可靠的系统,不仅能提高胰腺癌检测的准确性,还能在其他恶性肿瘤的早期诊断中得到更广泛的应用。
{"title":"Deep learning-based aggregate analysis to identify cut-off points for decision-making in pancreatic cancer detection","authors":"Gintautas Dzemyda,&nbsp;Olga Kurasova,&nbsp;Viktor Medvedev,&nbsp;Aušra Šubonienė,&nbsp;Aistė Gulla,&nbsp;Artūras Samuilis,&nbsp;Džiugas Jagminas,&nbsp;Kęstutis Strupas","doi":"10.1111/exsy.13614","DOIUrl":"10.1111/exsy.13614","url":null,"abstract":"<p>This study addresses the problem of detecting pancreatic cancer by classifying computed tomography (CT) images into cancerous and non-cancerous classes using the proposed deep learning-based aggregate analysis framework. The application of deep learning, as a branch of machine learning and artificial intelligence, to specific medical challenges can lead to the early detection of diseases, thus accelerating the process towards timely and effective intervention. The concept of classification is to reasonably select an optimal cut-off point, which is used as a threshold for evaluating the model results. The choice of this point is key to ensure efficient evaluation of the classification results, which directly affects the diagnostic accuracy. A significant aspect of this research is the incorporation of private CT images from Vilnius University Hospital Santaros Klinikos, combined with publicly available data sets. To investigate the capabilities of the deep learning-based framework and to maximize pancreatic cancer diagnostic performance, experimental studies were carried out combining data from different sources. Classification accuracy metrics such as the Youden index, (0, 1)-criterion, Matthew's correlation coefficient, the F1 score, LR+, LR−, balanced accuracy, and g-mean were used to find the optimal cut-off point in order to balance sensitivity and specificity. By carefully analyzing and comparing the obtained results, we aim to develop a reliable system that will not only improve the accuracy of pancreatic cancer detection but also have wider application in the early diagnosis of other malignancies.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"42 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140797763","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prediction of Liaoning province steel import and export trade based on deep learning models 基于深度学习模型的辽宁省钢铁进出口贸易预测
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-04-25 DOI: 10.1111/exsy.13615
Limin Zhang

In the field of deep learning, time series forecasting, particularly for economic and trade data, is a critical area of research. This study introduces a hybrid of auto regressive integrated moving average and gated recurrent unit (ARIMA-GRU) to enhance the prediction of steel import and export trade in Liaoning Province, addressing the limitations of traditional time series methods. Traditional models like ARIMA excel with linear data but often struggle with non-linear patterns and long-term dependencies. The ARIMA-GRU model combines ARIMA's linear data analysis with GRU's proficiency in non-linear pattern recognition, effectively capturing complex dynamics in economic datasets. Our experiments show that this hybrid approach surpasses traditional models in accuracy and reliability for forecasting steel trade, providing valuable insights for economic planning and strategic decision-making. This innovative approach not only advances the field of economic forecasting but also demonstrates the potential of integrating deep learning techniques in complex data analysis.

在深度学习领域,时间序列预测,尤其是经济和贸易数据的预测,是一个重要的研究领域。本研究针对传统时间序列方法的局限性,引入了自回归综合移动平均和门控循环单元(ARIMA-GRU)的混合模型,以加强对辽宁省钢铁进出口贸易的预测。ARIMA 等传统模型擅长处理线性数据,但在处理非线性模式和长期依赖关系时往往力不从心。ARIMA-GRU 模型结合了 ARIMA 的线性数据分析和 GRU 的非线性模式识别能力,能有效捕捉经济数据集中的复杂动态。我们的实验表明,这种混合方法在预测钢铁贸易的准确性和可靠性方面超越了传统模型,为经济规划和战略决策提供了宝贵的见解。这种创新方法不仅推动了经济预测领域的发展,还展示了在复杂数据分析中整合深度学习技术的潜力。
{"title":"Prediction of Liaoning province steel import and export trade based on deep learning models","authors":"Limin Zhang","doi":"10.1111/exsy.13615","DOIUrl":"10.1111/exsy.13615","url":null,"abstract":"<p>In the field of deep learning, time series forecasting, particularly for economic and trade data, is a critical area of research. This study introduces a hybrid of auto regressive integrated moving average and gated recurrent unit (ARIMA-GRU) to enhance the prediction of steel import and export trade in Liaoning Province, addressing the limitations of traditional time series methods. Traditional models like ARIMA excel with linear data but often struggle with non-linear patterns and long-term dependencies. The ARIMA-GRU model combines ARIMA's linear data analysis with GRU's proficiency in non-linear pattern recognition, effectively capturing complex dynamics in economic datasets. Our experiments show that this hybrid approach surpasses traditional models in accuracy and reliability for forecasting steel trade, providing valuable insights for economic planning and strategic decision-making. This innovative approach not only advances the field of economic forecasting but also demonstrates the potential of integrating deep learning techniques in complex data analysis.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"42 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140797825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modelling of healthcare data analytics using optimal machine learning model in big data environment 在大数据环境中使用最佳机器学习模型建立医疗数据分析模型
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-04-25 DOI: 10.1111/exsy.13612
Chelladurai Fancy, Nagappan Krishnaraj, K. Ishwarya, G. Raja, Shyamala Chandrasekaran

Recent advances in wireless networking, big data technologies, namely Internet of Things (IoT) 5G networks, health care big data analytics, and other technologies in artificial intelligence (AI) and wearables, have supported the progression of intellectual disease diagnosis methods. Medical data covers all patient data such as pharmacy texts, electronic health reports (EHR), prescriptions, study data from medical journals, clinical photographs, and diagnostic reports. Big data is a renowned method in the healthcare sector, with beneficial datasets that are highly difficult, voluminous, and rapid for healthcare providers for interpreting and computing using prevailing tools. This study combines concepts like deep learning (DL) and big data analytics in medical field. This article develops a new healthcare data analytics using optimal machine learning model in big data environment (HDAOML-BDE) technique. The presented HDAOML-BDE technique mainly aims to examine the healthcare data for disease detection and classification in the big data environment. For handling big data, the HDAOML-BDE technique uses Hadoop MapReduce environment. In addition, the HDAOML-BDE technique uses manta ray foraging optimization-based feature selection (MRFO-FS) technique to reduce high dimensionality problems. Moreover, the HDAOML-BDE method uses relevance vector machine (RVM) model for the healthcare data environment. Furthermore, the arithmetic optimization algorithm (AOA) is utilized for the parameter tuning of the RVM classifier. The simulation results of the HDAOML-BDE technique are tested on a healthcare dataset, and the outcomes portray the improved performance of the HDAOML-BDE strategy over recent approaches in different measures.

近年来,无线网络、大数据技术(即物联网(IoT)5G 网络)、医疗保健大数据分析以及人工智能(AI)和可穿戴设备等技术的发展,为智能疾病诊断方法的进步提供了支持。医疗数据涵盖所有患者数据,如药房文本、电子健康报告(EHR)、处方、医学期刊研究数据、临床照片和诊断报告。大数据是医疗保健领域的一种著名方法,其有益的数据集对于医疗保健提供者来说,使用现有工具进行解释和计算非常困难、庞大且快速。本研究将深度学习(DL)和大数据分析等概念结合到了医疗领域。本文利用大数据环境下的最优机器学习模型(HDAOML-BDE)技术开发了一种新的医疗数据分析方法。所提出的 HDAOML-BDE 技术主要是为了在大数据环境中对医疗数据进行疾病检测和分类。为了处理大数据,HDAOML-BDE 技术使用了 Hadoop MapReduce 环境。此外,HDAOML-BDE 技术还使用了基于蝠鲼觅食优化的特征选择(MRFO-FS)技术来减少高维问题。此外,HDAOML-BDE 方法还针对医疗数据环境使用了相关性向量机(RVM)模型。此外,还利用算术优化算法(AOA)来调整 RVM 分类器的参数。HDAOML-BDE 技术的仿真结果在医疗数据集上进行了测试,结果表明,HDAOML-BDE 策略在不同指标上的性能均优于近期采用的方法。
{"title":"Modelling of healthcare data analytics using optimal machine learning model in big data environment","authors":"Chelladurai Fancy,&nbsp;Nagappan Krishnaraj,&nbsp;K. Ishwarya,&nbsp;G. Raja,&nbsp;Shyamala Chandrasekaran","doi":"10.1111/exsy.13612","DOIUrl":"10.1111/exsy.13612","url":null,"abstract":"<p>Recent advances in wireless networking, big data technologies, namely Internet of Things (IoT) 5G networks, health care big data analytics, and other technologies in artificial intelligence (AI) and wearables, have supported the progression of intellectual disease diagnosis methods. Medical data covers all patient data such as pharmacy texts, electronic health reports (EHR), prescriptions, study data from medical journals, clinical photographs, and diagnostic reports. Big data is a renowned method in the healthcare sector, with beneficial datasets that are highly difficult, voluminous, and rapid for healthcare providers for interpreting and computing using prevailing tools. This study combines concepts like deep learning (DL) and big data analytics in medical field. This article develops a new healthcare data analytics using optimal machine learning model in big data environment (HDAOML-BDE) technique. The presented HDAOML-BDE technique mainly aims to examine the healthcare data for disease detection and classification in the big data environment. For handling big data, the HDAOML-BDE technique uses Hadoop MapReduce environment. In addition, the HDAOML-BDE technique uses manta ray foraging optimization-based feature selection (MRFO-FS) technique to reduce high dimensionality problems. Moreover, the HDAOML-BDE method uses relevance vector machine (RVM) model for the healthcare data environment. Furthermore, the arithmetic optimization algorithm (AOA) is utilized for the parameter tuning of the RVM classifier. The simulation results of the HDAOML-BDE technique are tested on a healthcare dataset, and the outcomes portray the improved performance of the HDAOML-BDE strategy over recent approaches in different measures.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"42 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140797827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimization on selecting XGBoost hyperparameters using meta-learning 利用元学习优化 XGBoost 超参数的选择
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-04-25 DOI: 10.1111/exsy.13611
Tiago Lima Marinho, Diego Carvalho do Nascimento, Bruno Almeida Pimentel

With computational evolution, there has been a growth in the number of machine learning algorithms and they became more complex and robust. A greater challenge is upon faster and more practical ways to find hyperparameters that will set up each algorithm individually. This article aims to use meta-learning as a practicable solution for recommending hyperparameters from similar datasets, through their meta-features structures, than to adopt the already trained XGBoost parameters for a new database. This reduced computational costs and also aimed to make real-time decision-making feasible or reduce any extra costs for companies for new information. The experimental results, adopting 198 data sets, attested to the success of the heuristics application using meta-learning to compare datasets structure analysis. Initially, a characterization of the datasets was performed by combining three groups of meta-features (general, statistical, and info-theory), so that there would be a way to compare the similarity between sets and, thus, apply meta-learning to recommend the hyperparameters. Later, the appropriate number of sets to characterize the XGBoost turning was tested. The obtained results were promising, showing an improved performance in the accuracy of the XGBoost, k = {4 − 6}, using the average of the hyperparameters values and, comparing to the standard grid-search hyperparameters set by default, it was obtained that, in 78.28% of the datasets, the meta-learning methodology performed better. This study, therefore, shows that the adoption of meta-learning is a competitive alternative to generalize the XGBoost model, expecting better statistics performance (accuracy etc.) rather than adjusting to a single/particular model.

随着计算技术的发展,机器学习算法的数量也在不断增加,而且变得更加复杂和强大。更大的挑战在于如何以更快、更实用的方式找到超参数,以单独设置每种算法。本文旨在使用元学习作为一种实用的解决方案,通过元特征结构从类似数据集中推荐超参数,而不是在新数据库中采用已经训练好的 XGBoost 参数。这不仅降低了计算成本,还旨在使实时决策变得可行,或减少公司获取新信息的额外成本。采用 198 个数据集的实验结果证明,使用元学习比较数据集结构分析的启发式应用是成功的。最初,通过结合三组元特征(一般特征、统计特征和信息理论特征)对数据集进行了特征描述,这样就有办法比较数据集之间的相似性,从而应用元学习推荐超参数。随后,测试了 XGBoost 转弯的特征集的适当数量。获得的结果很有希望,显示出使用超参数值的平均值(k = {4 - 6})提高了 XGBoost 的准确性,与默认设置的标准网格搜索超参数相比,在 78.28% 的数据集中,元学习方法的表现更好。因此,这项研究表明,采用元学习方法是推广 XGBoost 模型的一种有竞争力的替代方法,可望获得更好的统计性能(准确率等),而不是调整为单一/特定模型。
{"title":"Optimization on selecting XGBoost hyperparameters using meta-learning","authors":"Tiago Lima Marinho,&nbsp;Diego Carvalho do Nascimento,&nbsp;Bruno Almeida Pimentel","doi":"10.1111/exsy.13611","DOIUrl":"10.1111/exsy.13611","url":null,"abstract":"<p>With computational evolution, there has been a growth in the number of machine learning algorithms and they became more complex and robust. A greater challenge is upon faster and more practical ways to find hyperparameters that will set up each algorithm individually. This article aims to use meta-learning as a practicable solution for recommending hyperparameters from similar datasets, through their meta-features structures, than to adopt the already trained XGBoost parameters for a new database. This reduced computational costs and also aimed to make real-time decision-making feasible or reduce any extra costs for companies for new information. The experimental results, adopting 198 data sets, attested to the success of the heuristics application using meta-learning to compare datasets structure analysis. Initially, a characterization of the datasets was performed by combining three groups of meta-features (general, statistical, and info-theory), so that there would be a way to compare the similarity between sets and, thus, apply meta-learning to recommend the hyperparameters. Later, the appropriate number of sets to characterize the XGBoost turning was tested. The obtained results were promising, showing an improved performance in the accuracy of the XGBoost, <i>k</i> = {4 − 6}, using the average of the hyperparameters values and, comparing to the standard grid-search hyperparameters set by default, it was obtained that, in 78.28% of the datasets, the meta-learning methodology performed better. This study, therefore, shows that the adoption of meta-learning is a competitive alternative to generalize the XGBoost model, expecting better statistics performance (accuracy etc.) rather than adjusting to a single/particular model.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 9","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140797732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Expert Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1