首页 > 最新文献

Concurrency and Computation-Practice & Experience最新文献

英文 中文
Machine Learning-Based Data Deduplication: Techniques, Challenges, and Future Directions 基于机器学习的重复数据删除:技术、挑战和未来方向
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-25 DOI: 10.1002/cpe.70574
Ravneet Kaur, Harcharan Jit Singh, Inderveer Chana

Data deduplication plays an important role in modern data management as it reduces storage costs and ensures consistency by eliminating redundant records. The traditional data deduplication methods are effective for exact matches but struggle with adaptability and detecting near-exact duplicate records in unstructured or complex data. Machine learning (ML) addresses these limitations by using pattern recognition, feature learning, and statistical modeling to identify subtle similarities between records. This review classifies ML-based deduplication techniques into supervised, unsupervised, semi-supervised, and deep learning methodologies. It also discusses key challenges, including class imbalance, model interpretability, and computational overhead. The paper also explores recent developments in federated learning, real-time deduplication, and multimodal techniques to highlight current trends in these areas. Finally, the paper identifies key open issues and proposes a unified perspective for scalable, real-time deduplication systems that can accommodate diverse data types, structures, and system requirements.

重复数据删除在现代数据管理中发挥着重要作用,它通过消除冗余记录来降低存储成本并确保一致性。传统的重复数据删除方法对精确匹配是有效的,但在非结构化或复杂数据中难以适应和检测接近精确的重复记录。机器学习(ML)通过使用模式识别、特征学习和统计建模来识别记录之间微妙的相似性,从而解决了这些限制。本文将基于ml的重复数据删除技术分为监督式、无监督式、半监督式和深度学习方法。它还讨论了关键挑战,包括类不平衡、模型可解释性和计算开销。本文还探讨了联邦学习、实时重复数据删除和多模态技术的最新发展,以突出这些领域的当前趋势。最后,本文确定了关键的开放问题,并提出了一个统一的视角,可扩展的,实时的重复数据删除系统,可以适应不同的数据类型,结构和系统需求。
{"title":"Machine Learning-Based Data Deduplication: Techniques, Challenges, and Future Directions","authors":"Ravneet Kaur,&nbsp;Harcharan Jit Singh,&nbsp;Inderveer Chana","doi":"10.1002/cpe.70574","DOIUrl":"10.1002/cpe.70574","url":null,"abstract":"<div>\u0000 \u0000 <p>Data deduplication plays an important role in modern data management as it reduces storage costs and ensures consistency by eliminating redundant records. The traditional data deduplication methods are effective for exact matches but struggle with adaptability and detecting near-exact duplicate records in unstructured or complex data. Machine learning (ML) addresses these limitations by using pattern recognition, feature learning, and statistical modeling to identify subtle similarities between records. This review classifies ML-based deduplication techniques into supervised, unsupervised, semi-supervised, and deep learning methodologies. It also discusses key challenges, including class imbalance, model interpretability, and computational overhead. The paper also explores recent developments in federated learning, real-time deduplication, and multimodal techniques to highlight current trends in these areas. Finally, the paper identifies key open issues and proposes a unified perspective for scalable, real-time deduplication systems that can accommodate diverse data types, structures, and system requirements.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146091412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Explainable Ensemble Machine Learning Method for Electric Vehicles Energy Consumption Rate Estimation 一种可解释的集成机器学习方法用于电动汽车能耗率估算
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-24 DOI: 10.1002/cpe.70571
Mohammed Zaid Ghawy, Shuyan Chen, Sajan Shaikh, Aamir Hussain, Rajasekhar Balasubramanian, Yongfeng Ma

The rapid adoption of electric vehicles (EVs) highlights the need for intelligent systems to improve energy efficiency and optimize driving range. Since energy consumption and driving range modeling are closely related, understanding the energy consumption (EC) of EVs can provide essential insights to drivers and reduce “range anxiety.” Previous studies have relied on traditional analytical and statistical methods, which lack the representativeness of influential factors and the interpretability of the model applied in EC modeling. To address this issue, we propose an explainable ensemble machine learning model to predict EC of EVs, considering the most important features and the factors that exhibit greater influence on EC. The Spritmonitor public real-world dataset is used for this study. First, data preprocessing is conducted before feeding data into the ensemble method. Second, the Energy Consumption Rate (ECR) was predicted using Gradient Boosting Regression Trees (GBRT). The proposed predictive framework demonstrates superior prediction accuracy compared to baseline models. GBRT achieved the highest R 2 (1 and 0.99 for training and testing, respectively) and the lowest MAE (0.08) and RMSE (0.16) compared to other models, including XGBoost, LightGBM, and CatBoost. Finally, SHAP (Shapley Additive exPlanations) analysis was applied to explain the proposed model and identify the most influential dynamics factors, including driving range, capacity, speed, state of charge (SOC), ambient temperature, road type, driving style, air conditioning, and heating usage. The results suggest that the proposed framework can effectively enhance the prediction of the EC of EVs and facilitates the analyze driving factors, thereby supporting intelligent trip planning, adaptive energy-aware management in transportation systems and provide insightful feedback to drivers.

电动汽车的快速普及凸显了对智能系统的需求,以提高能源效率和优化续驶里程。由于能源消耗和续驶里程模型密切相关,因此了解电动汽车的能源消耗(EC)可以为驾驶员提供必要的见解,并减少“续驶里程焦虑”。以往的研究主要依靠传统的分析和统计方法,缺乏影响因素的代表性和模型的可解释性。为了解决这个问题,我们提出了一个可解释的集成机器学习模型来预测电动汽车的EC,考虑了最重要的特征和对EC影响较大的因素。本研究使用Spritmonitor公开的真实世界数据集。首先,在将数据输入集成方法之前进行数据预处理。其次,利用梯度增强回归树(GBRT)对能源消耗率(ECR)进行预测。与基线模型相比,所提出的预测框架具有更高的预测精度。与其他模型(包括XGBoost、LightGBM和CatBoost)相比,GBRT的r2最高(训练和测试分别为1和0.99),MAE最低(0.08),RMSE最低(0.16)。最后,采用Shapley加性解释(Shapley Additive exPlanations)分析对模型进行了解释,并确定了影响最大的动力学因素,包括续驶里程、容量、速度、充电状态(SOC)、环境温度、道路类型、驾驶方式、空调和供暖使用情况。结果表明,该框架能够有效增强电动汽车EC的预测能力,促进驱动因素分析,从而支持交通系统的智能出行规划、自适应能源意识管理,并为驾驶员提供有洞察力的反馈。
{"title":"An Explainable Ensemble Machine Learning Method for Electric Vehicles Energy Consumption Rate Estimation","authors":"Mohammed Zaid Ghawy,&nbsp;Shuyan Chen,&nbsp;Sajan Shaikh,&nbsp;Aamir Hussain,&nbsp;Rajasekhar Balasubramanian,&nbsp;Yongfeng Ma","doi":"10.1002/cpe.70571","DOIUrl":"10.1002/cpe.70571","url":null,"abstract":"<div>\u0000 \u0000 <p>The rapid adoption of electric vehicles (EVs) highlights the need for intelligent systems to improve energy efficiency and optimize driving range. Since energy consumption and driving range modeling are closely related, understanding the energy consumption (EC) of EVs can provide essential insights to drivers and reduce “range anxiety.” Previous studies have relied on traditional analytical and statistical methods, which lack the representativeness of influential factors and the interpretability of the model applied in EC modeling. To address this issue, we propose an explainable ensemble machine learning model to predict EC of EVs, considering the most important features and the factors that exhibit greater influence on EC. The Spritmonitor public real-world dataset is used for this study. First, data preprocessing is conducted before feeding data into the ensemble method. Second, the Energy Consumption Rate (ECR) was predicted using Gradient Boosting Regression Trees (GBRT). The proposed predictive framework demonstrates superior prediction accuracy compared to baseline models. GBRT achieved the highest <i>R</i>\u0000 <sup>2</sup> (1 and 0.99 for training and testing, respectively) and the lowest MAE (0.08) and RMSE (0.16) compared to other models, including XGBoost, LightGBM, and CatBoost. Finally, SHAP (Shapley Additive exPlanations) analysis was applied to explain the proposed model and identify the most influential dynamics factors, including driving range, capacity, speed, state of charge (SOC), ambient temperature, road type, driving style, air conditioning, and heating usage. The results suggest that the proposed framework can effectively enhance the prediction of the EC of EVs and facilitates the analyze driving factors, thereby supporting intelligent trip planning, adaptive energy-aware management in transportation systems and provide insightful feedback to drivers.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146058101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Carbon Emission Prediction for Gas Power Plants Based on Deep Learning Under Small-Sample Conditions 小样本条件下基于深度学习的燃气电厂碳排放预测
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-24 DOI: 10.1002/cpe.70591
Xiaozhou Fan, Zhe Wang, Hanwen Bi, Ruiyang Wang

Accurate forecasting of carbon emissions from power generation enterprises is essential under China's dual-control policy. Although deep learning methods show strong potential, studies on their optimal configuration remain limited. This paper proposed a hybrid deep learning framework integrating a convolutional neural network (CNN), bidirectional long short-term memory (BiLSTM), and an attention mechanism for carbon emission prediction in natural gas power plants. The present study utilized two distinct optimization methodologies: a structured design strategy encompassing light, medium, and heavy configurations, while the other employed Bayesian optimization for hyperparameter tuning. The models were evaluated using 5-fold cross-validation on 619 operational samples from two 487.1-MW condensing units in a power plant in Hainan, China. The medium configuration achieved the best balance between accuracy, efficiency, and stability, with R2 = 0.9833, RMSE = 0.0342, and MAE = 0.0242. Under small-sample conditions, the structured design approach outperformed Bayesian optimization by 0.16% in accuracy while requiring only 7.42% of the training time. The proposed framework provides an efficient and interpretable reference for selecting deep learning architectures in small-sample industrial regression tasks and supports intelligent, low-carbon power generation applications.

在中国的双重管制政策下,准确预测发电企业的碳排放至关重要。尽管深度学习方法显示出强大的潜力,但对其最优配置的研究仍然有限。本文提出了一种融合卷积神经网络(CNN)、双向长短期记忆(BiLSTM)和注意机制的混合深度学习框架,用于天然气电厂碳排放预测。本研究使用了两种不同的优化方法:一种包含轻、中、重配置的结构化设计策略,而另一种采用贝叶斯优化进行超参数调优。对中国海南某电厂两台487.1 mw冷凝机组的619个运行样本进行了5倍交叉验证,对模型进行了评估。介质配置在准确性、效率和稳定性之间达到了最佳平衡,R2 = 0.9833, RMSE = 0.0342, MAE = 0.0242。在小样本条件下,结构化设计方法的准确率比贝叶斯优化方法高0.16%,而训练时间仅为贝叶斯优化方法的7.42%。该框架为在小样本工业回归任务中选择深度学习架构提供了有效且可解释的参考,并支持智能、低碳发电应用。
{"title":"Carbon Emission Prediction for Gas Power Plants Based on Deep Learning Under Small-Sample Conditions","authors":"Xiaozhou Fan,&nbsp;Zhe Wang,&nbsp;Hanwen Bi,&nbsp;Ruiyang Wang","doi":"10.1002/cpe.70591","DOIUrl":"10.1002/cpe.70591","url":null,"abstract":"<div>\u0000 \u0000 <p>Accurate forecasting of carbon emissions from power generation enterprises is essential under China's dual-control policy. Although deep learning methods show strong potential, studies on their optimal configuration remain limited. This paper proposed a hybrid deep learning framework integrating a convolutional neural network (CNN), bidirectional long short-term memory (BiLSTM), and an attention mechanism for carbon emission prediction in natural gas power plants. The present study utilized two distinct optimization methodologies: a structured design strategy encompassing light, medium, and heavy configurations, while the other employed Bayesian optimization for hyperparameter tuning. The models were evaluated using 5-fold cross-validation on 619 operational samples from two 487.1-MW condensing units in a power plant in Hainan, China. The medium configuration achieved the best balance between accuracy, efficiency, and stability, with R<sup>2</sup> = 0.9833, RMSE = 0.0342, and MAE = 0.0242. Under small-sample conditions, the structured design approach outperformed Bayesian optimization by 0.16% in accuracy while requiring only 7.42% of the training time. The proposed framework provides an efficient and interpretable reference for selecting deep learning architectures in small-sample industrial regression tasks and supports intelligent, low-carbon power generation applications.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146091492","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Real-Time Automated Library Inventory System Based on Edge-Cloud Collaboration 基于边缘云协作的实时自动化图书馆库存系统
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-24 DOI: 10.1002/cpe.70573
Lu Zhu, Zhihui Gu, Kai Zhu, Xingcheng Xu, Jingzhi Wang, Yuanyuan Liu

Library inventory is vital for collection management and reader satisfaction. Conventional manual methods cannot support real-time updates, while existing automated solutions relying on centralized cloud computing suffer from bandwidth and latency limitations. To address these issues, we propose an edge-cloud collaborative real-time book inventory system. Spine detection and text recognition are executed on embedded edge devices, while the cloud handles rapid data retrieval to balance timeliness and accuracy. We design lightweight models for edge deployment, including the Library You Only Look Once (Lib-YOLO) detector with a StarNet backbone, shared convolutional head, and dual-scale hierarchical detection, supporting rotated objects for robust spine extraction. The optimized paddle practical optical character recognition (PP-OCR) pipeline removes text rectification and integrates a filtering algorithm to reduce redundant computation and improve efficiency. Deployed on an NVIDIA Jetson Nano, the system achieves 73 ms spine detection latency, 191 ms text recognition latency, and 97.1% overall accuracy under simulated library conditions. The Lib-YOLO model contains only 1.39 M parameters with 99% mean average precision (mAP), demonstrating the feasibility of precise, real-time inventorying in resource-constrained embedded environments.

图书馆库存对馆藏管理和读者满意度至关重要。传统的手动方法无法支持实时更新,而依赖集中式云计算的现有自动化解决方案则受到带宽和延迟的限制。为了解决这些问题,我们提出了一个边缘云协作实时图书库存系统。脊柱检测和文本识别在嵌入式边缘设备上执行,而云处理快速数据检索以平衡及时性和准确性。我们为边缘部署设计了轻量级模型,包括带有StarNet主干的Library You Only Look Once (Lib-YOLO)检测器,共享卷积头和双尺度分层检测,支持旋转对象进行鲁棒脊柱提取。优化后的桨形实用光学字符识别(PP-OCR)管道去除了文本纠错,并集成了滤波算法,减少了冗余计算,提高了效率。该系统部署在NVIDIA Jetson Nano上,在模拟图书馆条件下实现了73毫秒的脊柱检测延迟,191毫秒的文本识别延迟和97.1%的总体准确率。Lib-YOLO模型仅包含1.39 M个参数,平均平均精度(mAP)为99%,证明了在资源受限的嵌入式环境中实现精确、实时库存的可行性。
{"title":"A Real-Time Automated Library Inventory System Based on Edge-Cloud Collaboration","authors":"Lu Zhu,&nbsp;Zhihui Gu,&nbsp;Kai Zhu,&nbsp;Xingcheng Xu,&nbsp;Jingzhi Wang,&nbsp;Yuanyuan Liu","doi":"10.1002/cpe.70573","DOIUrl":"10.1002/cpe.70573","url":null,"abstract":"<div>\u0000 \u0000 <p>Library inventory is vital for collection management and reader satisfaction. Conventional manual methods cannot support real-time updates, while existing automated solutions relying on centralized cloud computing suffer from bandwidth and latency limitations. To address these issues, we propose an edge-cloud collaborative real-time book inventory system. Spine detection and text recognition are executed on embedded edge devices, while the cloud handles rapid data retrieval to balance timeliness and accuracy. We design lightweight models for edge deployment, including the Library You Only Look Once (Lib-YOLO) detector with a StarNet backbone, shared convolutional head, and dual-scale hierarchical detection, supporting rotated objects for robust spine extraction. The optimized paddle practical optical character recognition (PP-OCR) pipeline removes text rectification and integrates a filtering algorithm to reduce redundant computation and improve efficiency. Deployed on an NVIDIA Jetson Nano, the system achieves 73 ms spine detection latency, 191 ms text recognition latency, and 97.1% overall accuracy under simulated library conditions. The Lib-YOLO model contains only 1.39 M parameters with 99% mean average precision (mAP), demonstrating the feasibility of precise, real-time inventorying in resource-constrained embedded environments.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146057953","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SecureChain: A Blockchain-Based Secure Model for Sharing Privacy-Preserved Data Using Local Differential Privacy SecureChain:一种基于区块链的安全模型,用于使用本地差分隐私共享隐私保护数据
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-24 DOI: 10.1002/cpe.70473
Altaf Hussain, Laraib Javed, Muhammad Inam Ul Haq, Razaullah Khan, Wajahat Akbar, Razaz Waheeb Attar, Ahmed Alhazmi, Amal Hassan Alhazmi, Tariq Hussain

Privacy-Preserving Data Sharing (PPDS) masks the individual's collected data (e.g., medical healthcare data) before being disseminated by organizations for analysis and research. Patient data contains sensitive values that must be dealt with while ensuring certain privacy conditions are met. This minimizes the risk of re-identification of an individual record from the group of privacy-preserved data. However, with the advancement in technology (i.e., Big Data, the Internet of Things (IoT), and Blockchain), the existing classical privacy-preserving techniques are becoming obsolete. In this paper, we propose a blockchain-based secure data sharing technique named “SecureChain”, which preserves the privacy of an individual record using local differential privacy (LDP). The three distinguished features of the proposed approach are lower latency, higher throughput, and improved privacy. The proposed model outperforms the benchmarks in terms of both latency and throughput. In terms of precision, the proposed method improves the accuracy to 88.53% compared to its counterparts, which achieved 49% and 85% accuracy. The results of the experiment verify that the proposed approach outperforms its counterparts.

保护隐私的数据共享(PPDS)在组织传播用于分析和研究之前掩盖个人收集的数据(例如医疗保健数据)。患者数据包含必须处理的敏感值,同时确保满足某些隐私条件。这最大限度地减少了从一组隐私保护数据中重新识别单个记录的风险。然而,随着技术的进步(即大数据、物联网(IoT)和区块链),现有的经典隐私保护技术已经过时。在本文中,我们提出了一种名为“SecureChain”的基于区块链的安全数据共享技术,该技术使用本地差分隐私(LDP)来保护单个记录的隐私。该方法的三个显著特征是低延迟、高吞吐量和改进的隐私性。所建议的模型在延迟和吞吐量方面都优于基准测试。在精度方面,与同类方法相比,本文方法的准确率达到了49%和85%,提高了88.53%。实验结果验证了该方法的有效性。
{"title":"SecureChain: A Blockchain-Based Secure Model for Sharing Privacy-Preserved Data Using Local Differential Privacy","authors":"Altaf Hussain,&nbsp;Laraib Javed,&nbsp;Muhammad Inam Ul Haq,&nbsp;Razaullah Khan,&nbsp;Wajahat Akbar,&nbsp;Razaz Waheeb Attar,&nbsp;Ahmed Alhazmi,&nbsp;Amal Hassan Alhazmi,&nbsp;Tariq Hussain","doi":"10.1002/cpe.70473","DOIUrl":"10.1002/cpe.70473","url":null,"abstract":"<div>\u0000 \u0000 <p>Privacy-Preserving Data Sharing (PPDS) masks the individual's collected data (e.g., medical healthcare data) before being disseminated by organizations for analysis and research. Patient data contains sensitive values that must be dealt with while ensuring certain privacy conditions are met. This minimizes the risk of re-identification of an individual record from the group of privacy-preserved data. However, with the advancement in technology (i.e., Big Data, the Internet of Things (IoT), and Blockchain), the existing classical privacy-preserving techniques are becoming obsolete. In this paper, we propose a blockchain-based secure data sharing technique named “SecureChain”, which preserves the privacy of an individual record using local differential privacy (LDP). The three distinguished features of the proposed approach are lower latency, higher throughput, and improved privacy. The proposed model outperforms the benchmarks in terms of both latency and throughput. In terms of precision, the proposed method improves the accuracy to 88.53% compared to its counterparts, which achieved 49% and 85% accuracy. The results of the experiment verify that the proposed approach outperforms its counterparts.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146058102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AT-SPNet: A Personalized Federated Spatio-Temporal Modeling Method for Cross-City Traffic Prediction 面向跨城市交通预测的个性化联邦时空建模方法
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-23 DOI: 10.1002/cpe.70577
Ying Wang, Renjie Fan, Bo Gong, Hong Wen, Yuanxi Yu

For cross-city traffic prediction, the significant heterogeneity of traffic data across cities and the requirement for privacy protection make it challenging for conventional centralized spatiotemporal graph modeling techniques to balance predictive performance and data security. Therefore, this paper proposes AT-SPNet, a personalized federated spatiotemporal modeling approach specifically designed for cross-city traffic prediction. This method decouples the spatiotemporal modeling paths through the construction of a shared temporal branch and a hidden local spatial branch, thereby mitigating the heterogeneity of cross-city traffic data while preserving privacy. In the temporal branch, Gated Recurrent Units and a multi-head attention mechanism are incorporated to capture temporal dependencies, and a Squeeze-and-Excitation module is employed to enhance the extraction of informative features. In the spatial branch, a Spatial Attention Fusion module based on a triple-attention mechanism is designed to capture spatial features from multiple spatial perspectives, combined with static graph convolution and dynamic graph attention to construct a dual-modal information fusion path. Furthermore, to alleviate the adverse effects of cross-city data heterogeneity in federated training, a personalized federated learning strategy is introduced, which enables differentiated fusion of client spatial features without sharing raw data. Experiments on four real-world traffic datasets demonstrate that AT-SPNet outperforms existing methods in both prediction accuracy and cross-city generalization, validating the effectiveness and practical applicability of the proposed approach for cross-city traffic prediction.

对于跨城市交通预测,由于城市间交通数据的显著异质性和对隐私保护的要求,使得传统的集中式时空图建模技术难以平衡预测性能和数据安全性。为此,本文提出了一种专门为跨城市交通预测设计的个性化联邦时空建模方法AT-SPNet。该方法通过构建一个共享的时间分支和一个隐藏的局部空间分支来解耦时空建模路径,从而在保护隐私的同时减轻了跨城市交通数据的异质性。在时间分支中,采用门控循环单元和多头注意机制来捕获时间依赖性,并采用挤压和激励模块来增强信息特征的提取。在空间分支中,设计了基于三注意机制的空间注意融合模块,从多个空间视角捕捉空间特征,结合静态图卷积和动态图注意构建双模态信息融合路径。此外,为了缓解跨城市数据异构对联邦训练的不利影响,提出了一种个性化的联邦学习策略,在不共享原始数据的情况下实现客户端空间特征的差异化融合。在4个真实交通数据集上的实验表明,AT-SPNet在预测精度和跨城市泛化方面都优于现有方法,验证了该方法在跨城市交通预测中的有效性和实用性。
{"title":"AT-SPNet: A Personalized Federated Spatio-Temporal Modeling Method for Cross-City Traffic Prediction","authors":"Ying Wang,&nbsp;Renjie Fan,&nbsp;Bo Gong,&nbsp;Hong Wen,&nbsp;Yuanxi Yu","doi":"10.1002/cpe.70577","DOIUrl":"https://doi.org/10.1002/cpe.70577","url":null,"abstract":"<div>\u0000 \u0000 <p>For cross-city traffic prediction, the significant heterogeneity of traffic data across cities and the requirement for privacy protection make it challenging for conventional centralized spatiotemporal graph modeling techniques to balance predictive performance and data security. Therefore, this paper proposes AT-SPNet, a personalized federated spatiotemporal modeling approach specifically designed for cross-city traffic prediction. This method decouples the spatiotemporal modeling paths through the construction of a shared temporal branch and a hidden local spatial branch, thereby mitigating the heterogeneity of cross-city traffic data while preserving privacy. In the temporal branch, Gated Recurrent Units and a multi-head attention mechanism are incorporated to capture temporal dependencies, and a Squeeze-and-Excitation module is employed to enhance the extraction of informative features. In the spatial branch, a Spatial Attention Fusion module based on a triple-attention mechanism is designed to capture spatial features from multiple spatial perspectives, combined with static graph convolution and dynamic graph attention to construct a dual-modal information fusion path. Furthermore, to alleviate the adverse effects of cross-city data heterogeneity in federated training, a personalized federated learning strategy is introduced, which enables differentiated fusion of client spatial features without sharing raw data. Experiments on four real-world traffic datasets demonstrate that AT-SPNet outperforms existing methods in both prediction accuracy and cross-city generalization, validating the effectiveness and practical applicability of the proposed approach for cross-city traffic prediction.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146091393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Performance and Cost Evaluation of StarPU on AWS: Case Studies With Dense Linear Algebra Kernels and N-Body Simulations 基于AWS的StarPU性能和成本评估:基于密集线性代数核和n体模拟的案例研究
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-23 DOI: 10.1002/cpe.70582
Vanderlei Munhoz, Vinicius G. Pinto, João V. F. Lima, Márcio Castro, Daniel Cordeiro, Emilio Francesquini

Task-based programming interfaces introduce a paradigm in which computations are decomposed into fine-grained units of work known as “tasks”. StarPU is a runtime system originally developed to support task-based parallelism on on-premise heterogeneous architectures by abstracting low-level hardware details and efficiently managing resource scheduling. It enables developers to express applications as task graphs with explicit data dependencies, which are then dynamically scheduled across available processing units, such as CPUs and GPUs. In recent years, major cloud providers have begun offering virtual machines equipped with both CPUs and GPUs, allowing researchers to deploy and execute parallel workloads in virtual heterogeneous clusters. However, the performance and cost effectiveness of executing StarPU-based applications in public cloud environments remain unclear, particularly due to variability in hardware configurations, network performance, ever-changing pricing models, and computing performance due to virtualization and multi-tenancy. In this paper, we evaluate the performance and cost-efficiency of StarPU on Amazon Elastic Compute Cloud (EC2) using dense linear algebra kernels and N-Body simulations as case studies. Our experiments consider different cluster configurations, including powerful and more expensive instances with four NVIDIA GPUs per node (which we refer to as “fat nodes”), and less powerful and lower-cost instances with a single NVIDIA GPU per node (which we refer to as “thin nodes”). Our results show that arithmetic precision affects the performance–cost trade-off for dense linear algebra applications, whereas N-Body simulations consistently achieve better cost-efficiency on thin-node clusters. These findings underscore the challenges of optimizing HPC workloads for performance and cost in cloud environments.

基于任务的编程接口引入了一种范式,在这种范式中,计算被分解为称为“任务”的细粒度工作单元。StarPU是一个运行时系统,最初是通过抽象底层硬件细节和有效管理资源调度来支持基于任务的本地异构架构的并行性。它使开发人员能够将应用程序表示为具有显式数据依赖关系的任务图,然后在可用的处理单元(如cpu和gpu)之间动态调度。近年来,主要的云提供商已经开始提供配备cpu和gpu的虚拟机,允许研究人员在虚拟异构集群中部署和执行并行工作负载。然而,在公共云环境中执行基于starpu的应用程序的性能和成本效益仍然不清楚,特别是由于硬件配置、网络性能、不断变化的定价模型以及虚拟化和多租户导致的计算性能的可变性。在本文中,我们使用密集线性代数核和N-Body模拟作为案例研究,评估了StarPU在Amazon Elastic Compute Cloud (EC2)上的性能和成本效率。我们的实验考虑了不同的集群配置,包括每个节点具有四个NVIDIA GPU的强大且更昂贵的实例(我们称之为“胖节点”),以及每个节点具有单个NVIDIA GPU的功能较弱且成本较低的实例(我们称之为“瘦节点”)。我们的研究结果表明,算法精度影响密集线性代数应用的性能成本权衡,而N-Body模拟在瘦节点集群上始终获得更好的成本效益。这些发现强调了在云环境中优化高性能计算工作负载的性能和成本所面临的挑战。
{"title":"Performance and Cost Evaluation of StarPU on AWS: Case Studies With Dense Linear Algebra Kernels and N-Body Simulations","authors":"Vanderlei Munhoz,&nbsp;Vinicius G. Pinto,&nbsp;João V. F. Lima,&nbsp;Márcio Castro,&nbsp;Daniel Cordeiro,&nbsp;Emilio Francesquini","doi":"10.1002/cpe.70582","DOIUrl":"10.1002/cpe.70582","url":null,"abstract":"<p>Task-based programming interfaces introduce a paradigm in which computations are decomposed into fine-grained units of work known as “tasks”. StarPU is a runtime system originally developed to support task-based parallelism on on-premise heterogeneous architectures by abstracting low-level hardware details and efficiently managing resource scheduling. It enables developers to express applications as task graphs with explicit data dependencies, which are then dynamically scheduled across available processing units, such as CPUs and GPUs. In recent years, major cloud providers have begun offering virtual machines equipped with both CPUs and GPUs, allowing researchers to deploy and execute parallel workloads in virtual heterogeneous clusters. However, the performance and cost effectiveness of executing StarPU-based applications in public cloud environments remain unclear, particularly due to variability in hardware configurations, network performance, ever-changing pricing models, and computing performance due to virtualization and multi-tenancy. In this paper, we evaluate the performance and cost-efficiency of StarPU on Amazon Elastic Compute Cloud (EC2) using dense linear algebra kernels and N-Body simulations as case studies. Our experiments consider different cluster configurations, including powerful and more expensive instances with four NVIDIA GPUs per node (which we refer to as “fat nodes”), and less powerful and lower-cost instances with a single NVIDIA GPU per node (which we refer to as “thin nodes”). Our results show that arithmetic precision affects the performance–cost trade-off for dense linear algebra applications, whereas N-Body simulations consistently achieve better cost-efficiency on thin-node clusters. These findings underscore the challenges of optimizing HPC workloads for performance and cost in cloud environments.</p>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/cpe.70582","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146057779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Efficient Deep Learning Model for Multiclass Brain Tumor Classification Using MRI Images With Triple Explainability 基于三重可解释性MRI图像的脑肿瘤分类深度学习模型
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-23 DOI: 10.1002/cpe.70548
Kashif Mazhar, Pragya Dwivedi, Vibhor Kant

Brain tumors (BT) are considered a major health challenge around the world, which needs early detection; therefore, effective treatment strategies can be planned. This kind of cancer greatly diminishes the patient's quality and lifespan, which opens a gateway for early diagnosis and effective treatment. The medical professionals need assistance with this difficult and error-prone process; it is also mandatory to augment the interpretability and accuracy of the recognition model. To achieve such a goal, a hybrid deep learning model superior with explainable AI is introduced in the proposed framework, which performs brain tumor classification and model interpretation from MRI. The proposed study involves four key steps: Pre-processing, segmentation, classification, and analysis. The input images are initially pre-processed using a median-boosted Kuan Filtering (Me-KF) to remove any noise in the data and improve the subsequent segmentation procedure. After pre-processing, the Extended Multi-Inception Attention U-Net (ExMIA_U-Net) technique is added to effectively separate the brain tumor region. Finally, a deep learning method based on Convolution Attentive assisted EfficientNetB0 (CA-EfficientNetB0) is presented to categorize the many categories of brain tumors, comprising gliomas, meningiomas, pituitary tumors and normal tumors. This model uses Shapley additive explanation (SHAP), Local interpretable model-agnostic explanations (LIME), and Gradient-weighted Class Activation Mapping (Grad-CAM) for model interpretation. The proposed model uses a brain tumor classification dataset. In the results section, the proposed model is compared to many other prevailing schemes and it achieves 99.45% accuracy, 99.16% precision, 98.97% recall, and a 99.06% F1-score. The results show that an efficient, interpretable, robust and better model is developed for brain tumor classification.

脑肿瘤(BT)被认为是世界各地的一个重大健康挑战,需要及早发现;因此,可以制定有效的治疗策略。这类癌症大大降低了患者的生活质量和寿命,为早期诊断和有效治疗打开了大门。医疗专业人员在这一困难和容易出错的过程中需要帮助;增强识别模型的可解释性和准确性也是必须的。为了实现这一目标,在提出的框架中引入了具有可解释AI的混合深度学习模型,该模型可以从MRI中进行脑肿瘤分类和模型解释。该研究包括四个关键步骤:预处理、分割、分类和分析。输入图像最初使用中值增强宽滤波(Me-KF)进行预处理,以去除数据中的任何噪声并改进随后的分割过程。在预处理后,加入扩展多inception注意U-Net (ExMIA_U-Net)技术,有效分离脑肿瘤区域。最后,提出了一种基于卷积细心辅助高效netb0 (CA-EfficientNetB0)的深度学习方法,用于脑肿瘤的分类,包括胶质瘤、脑膜瘤、垂体瘤和正常肿瘤。该模型使用Shapley加性解释(SHAP)、局部可解释模型不可知解释(LIME)和梯度加权类激活映射(Grad-CAM)进行模型解释。该模型使用脑肿瘤分类数据集。在结果部分,将所提出的模型与许多其他流行方案进行了比较,结果表明,该模型达到了99.45%的准确率、99.16%的精度、98.97%的召回率和99.06%的f1分数。结果表明,建立了一种高效、可解释、鲁棒性较好的脑肿瘤分类模型。
{"title":"An Efficient Deep Learning Model for Multiclass Brain Tumor Classification Using MRI Images With Triple Explainability","authors":"Kashif Mazhar,&nbsp;Pragya Dwivedi,&nbsp;Vibhor Kant","doi":"10.1002/cpe.70548","DOIUrl":"https://doi.org/10.1002/cpe.70548","url":null,"abstract":"<div>\u0000 \u0000 <p>Brain tumors (BT) are considered a major health challenge around the world, which needs early detection; therefore, effective treatment strategies can be planned. This kind of cancer greatly diminishes the patient's quality and lifespan, which opens a gateway for early diagnosis and effective treatment. The medical professionals need assistance with this difficult and error-prone process; it is also mandatory to augment the interpretability and accuracy of the recognition model. To achieve such a goal, a hybrid deep learning model superior with explainable AI is introduced in the proposed framework, which performs brain tumor classification and model interpretation from MRI. The proposed study involves four key steps: Pre-processing, segmentation, classification, and analysis. The input images are initially pre-processed using a median-boosted Kuan Filtering (Me-KF) to remove any noise in the data and improve the subsequent segmentation procedure. After pre-processing, the Extended Multi-Inception Attention U-Net (ExMIA_U-Net) technique is added to effectively separate the brain tumor region. Finally, a deep learning method based on Convolution Attentive assisted EfficientNetB0 (CA-EfficientNetB0) is presented to categorize the many categories of brain tumors, comprising gliomas, meningiomas, pituitary tumors and normal tumors. This model uses Shapley additive explanation (SHAP), Local interpretable model-agnostic explanations (LIME), and Gradient-weighted Class Activation Mapping (Grad-CAM) for model interpretation. The proposed model uses a brain tumor classification dataset. In the results section, the proposed model is compared to many other prevailing schemes and it achieves 99.45% accuracy, 99.16% precision, 98.97% recall, and a 99.06% F1-score. The results show that an efficient, interpretable, robust and better model is developed for brain tumor classification.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146091390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MADS-UC: Finding Key Users in Online Social Networks Through Users Activation Activeness and Combination Weighting MCDM MADS-UC:基于用户激活活跃度和组合加权MCDM的在线社交网络关键用户挖掘
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-23 DOI: 10.1002/cpe.70566
Pingle Yang, Laijun Zhao, Fanyuan Meng, Huiyong Li, Lixin Zhou, Chen Dong

In online social networks, identifying influential users is crucial for maintaining network stability and accelerating information dissemination. However, most of the existing research evaluate the influence of users according to the topological structure of local or global networks, which ignores the historical information and social interactions of users. In this work, we introduce a novel algorithm called MADS-UC to pinpoint the influential super-spreaders of information. Specifically, three kinds of activation influence between each pair of users are defined to better portray the mutual interactions with each other, and the proposed measure makes full consideration of the topological and historical information of networks, and the social interactions of users. Then, a hybrid multi-attribute decision-making algorithm is put forward to evaluate the influence of users, while the obtained three kinds of activation influence are introduced as basic indicators, and the weight for each indicator is determined from both subjective and objective dimensions. Finally, to achieve a good balance between algorithm accuracy and time complexity, the influential users are identified by considering the interactions between a user and other users within its influence range. We collect seven real-world datasets from the Sina Weibo platform in 2024 about mobile product Honor, and a series of experiments is conducted to validate the effectiveness of the MADS-UC algorithm. Experimental results show that MADS-UC surpasses six widely used algorithms in robustness, sensitivity analysis, and distinguishing capability, which is useful in accelerating the dissemination of information in the product marketing process.

在在线社交网络中,识别有影响力的用户对于维护网络稳定和加速信息传播至关重要。然而,现有的研究大多是根据局部或全局网络的拓扑结构来评估用户的影响,而忽略了用户的历史信息和社会互动。在这项工作中,我们引入了一种称为MADS-UC的新算法来确定有影响力的信息超级传播者。具体而言,定义了每对用户之间的三种激活影响,以更好地描述彼此之间的交互,并且所提出的度量充分考虑了网络的拓扑和历史信息以及用户的社会交互。然后,提出了一种混合多属性决策算法来评价用户的影响,并将得到的三种激活影响作为基本指标,从主观和客观两个维度确定各指标的权重。最后,为了在算法精度和时间复杂度之间取得良好的平衡,通过考虑用户与其影响范围内其他用户之间的交互来识别有影响力的用户。我们从2024年的新浪微博平台上收集了7个关于移动产品荣耀的真实数据集,并进行了一系列实验来验证MADS-UC算法的有效性。实验结果表明,MADS-UC算法在鲁棒性、灵敏度分析和区分能力等方面均优于6种常用算法,有助于加快产品营销过程中的信息传播。
{"title":"MADS-UC: Finding Key Users in Online Social Networks Through Users Activation Activeness and Combination Weighting MCDM","authors":"Pingle Yang,&nbsp;Laijun Zhao,&nbsp;Fanyuan Meng,&nbsp;Huiyong Li,&nbsp;Lixin Zhou,&nbsp;Chen Dong","doi":"10.1002/cpe.70566","DOIUrl":"10.1002/cpe.70566","url":null,"abstract":"<div>\u0000 \u0000 <p>In online social networks, identifying influential users is crucial for maintaining network stability and accelerating information dissemination. However, most of the existing research evaluate the influence of users according to the topological structure of local or global networks, which ignores the historical information and social interactions of users. In this work, we introduce a novel algorithm called MADS-UC to pinpoint the influential super-spreaders of information. Specifically, three kinds of activation influence between each pair of users are defined to better portray the mutual interactions with each other, and the proposed measure makes full consideration of the topological and historical information of networks, and the social interactions of users. Then, a hybrid multi-attribute decision-making algorithm is put forward to evaluate the influence of users, while the obtained three kinds of activation influence are introduced as basic indicators, and the weight for each indicator is determined from both subjective and objective dimensions. Finally, to achieve a good balance between algorithm accuracy and time complexity, the influential users are identified by considering the interactions between a user and other users within its influence range. We collect seven real-world datasets from the Sina Weibo platform in 2024 about mobile product Honor, and a series of experiments is conducted to validate the effectiveness of the MADS-UC algorithm. Experimental results show that MADS-UC surpasses six widely used algorithms in robustness, sensitivity analysis, and distinguishing capability, which is useful in accelerating the dissemination of information in the product marketing process.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146058063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Efficient Deep Learning Model for Multiclass Brain Tumor Classification Using MRI Images With Triple Explainability 基于三重可解释性MRI图像的脑肿瘤分类深度学习模型
IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-23 DOI: 10.1002/cpe.70548
Kashif Mazhar, Pragya Dwivedi, Vibhor Kant

Brain tumors (BT) are considered a major health challenge around the world, which needs early detection; therefore, effective treatment strategies can be planned. This kind of cancer greatly diminishes the patient's quality and lifespan, which opens a gateway for early diagnosis and effective treatment. The medical professionals need assistance with this difficult and error-prone process; it is also mandatory to augment the interpretability and accuracy of the recognition model. To achieve such a goal, a hybrid deep learning model superior with explainable AI is introduced in the proposed framework, which performs brain tumor classification and model interpretation from MRI. The proposed study involves four key steps: Pre-processing, segmentation, classification, and analysis. The input images are initially pre-processed using a median-boosted Kuan Filtering (Me-KF) to remove any noise in the data and improve the subsequent segmentation procedure. After pre-processing, the Extended Multi-Inception Attention U-Net (ExMIA_U-Net) technique is added to effectively separate the brain tumor region. Finally, a deep learning method based on Convolution Attentive assisted EfficientNetB0 (CA-EfficientNetB0) is presented to categorize the many categories of brain tumors, comprising gliomas, meningiomas, pituitary tumors and normal tumors. This model uses Shapley additive explanation (SHAP), Local interpretable model-agnostic explanations (LIME), and Gradient-weighted Class Activation Mapping (Grad-CAM) for model interpretation. The proposed model uses a brain tumor classification dataset. In the results section, the proposed model is compared to many other prevailing schemes and it achieves 99.45% accuracy, 99.16% precision, 98.97% recall, and a 99.06% F1-score. The results show that an efficient, interpretable, robust and better model is developed for brain tumor classification.

脑肿瘤(BT)被认为是世界各地的一个重大健康挑战,需要及早发现;因此,可以制定有效的治疗策略。这类癌症大大降低了患者的生活质量和寿命,为早期诊断和有效治疗打开了大门。医疗专业人员在这一困难和容易出错的过程中需要帮助;增强识别模型的可解释性和准确性也是必须的。为了实现这一目标,在提出的框架中引入了具有可解释AI的混合深度学习模型,该模型可以从MRI中进行脑肿瘤分类和模型解释。该研究包括四个关键步骤:预处理、分割、分类和分析。输入图像最初使用中值增强宽滤波(Me-KF)进行预处理,以去除数据中的任何噪声并改进随后的分割过程。在预处理后,加入扩展多inception注意U-Net (ExMIA_U-Net)技术,有效分离脑肿瘤区域。最后,提出了一种基于卷积细心辅助高效netb0 (CA-EfficientNetB0)的深度学习方法,用于脑肿瘤的分类,包括胶质瘤、脑膜瘤、垂体瘤和正常肿瘤。该模型使用Shapley加性解释(SHAP)、局部可解释模型不可知解释(LIME)和梯度加权类激活映射(Grad-CAM)进行模型解释。该模型使用脑肿瘤分类数据集。在结果部分,将所提出的模型与许多其他流行方案进行了比较,结果表明,该模型达到了99.45%的准确率、99.16%的精度、98.97%的召回率和99.06%的f1分数。结果表明,建立了一种高效、可解释、鲁棒性较好的脑肿瘤分类模型。
{"title":"An Efficient Deep Learning Model for Multiclass Brain Tumor Classification Using MRI Images With Triple Explainability","authors":"Kashif Mazhar,&nbsp;Pragya Dwivedi,&nbsp;Vibhor Kant","doi":"10.1002/cpe.70548","DOIUrl":"https://doi.org/10.1002/cpe.70548","url":null,"abstract":"<div>\u0000 \u0000 <p>Brain tumors (BT) are considered a major health challenge around the world, which needs early detection; therefore, effective treatment strategies can be planned. This kind of cancer greatly diminishes the patient's quality and lifespan, which opens a gateway for early diagnosis and effective treatment. The medical professionals need assistance with this difficult and error-prone process; it is also mandatory to augment the interpretability and accuracy of the recognition model. To achieve such a goal, a hybrid deep learning model superior with explainable AI is introduced in the proposed framework, which performs brain tumor classification and model interpretation from MRI. The proposed study involves four key steps: Pre-processing, segmentation, classification, and analysis. The input images are initially pre-processed using a median-boosted Kuan Filtering (Me-KF) to remove any noise in the data and improve the subsequent segmentation procedure. After pre-processing, the Extended Multi-Inception Attention U-Net (ExMIA_U-Net) technique is added to effectively separate the brain tumor region. Finally, a deep learning method based on Convolution Attentive assisted EfficientNetB0 (CA-EfficientNetB0) is presented to categorize the many categories of brain tumors, comprising gliomas, meningiomas, pituitary tumors and normal tumors. This model uses Shapley additive explanation (SHAP), Local interpretable model-agnostic explanations (LIME), and Gradient-weighted Class Activation Mapping (Grad-CAM) for model interpretation. The proposed model uses a brain tumor classification dataset. In the results section, the proposed model is compared to many other prevailing schemes and it achieves 99.45% accuracy, 99.16% precision, 98.97% recall, and a 99.06% F1-score. The results show that an efficient, interpretable, robust and better model is developed for brain tumor classification.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"38 3","pages":""},"PeriodicalIF":1.5,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146091381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Concurrency and Computation-Practice & Experience
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1