首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
HAI-Net: Skin Lesion Segmentation Using a High-Performance Adaptive Attention and Information Interaction Network HAI-Net:基于高性能自适应注意与信息交互网络的皮肤病变分割
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-10 DOI: 10.1002/ima.70266
Chao Fan, Li Chen, Mengyang Yun, Huijun Zhao, Bincheng Peng

Skin lesion segmentation from dermoscopic images must be done accurately and consistently in order to diagnose diseases and arrange treatments. However, when dealing with issues like fuzzy lesion region boundaries, multiscale features, and notable variations in the lesion region's size, shape, and color, existing methods typically have high computational complexity and large parameter counts. They also frequently suffer from decreased segmentation accuracy due to inadequate capture of local features and global information. In this paper, a lightweight deep learning network based on high-performance adaptive attention is proposed to overcome these issues. Specifically, a deep convolutional neural network is introduced to capture local information. Meanwhile, we create a high-performance adaptive attention feature fusion module (EAAF) that uses dynamic feature selection to achieve adaptive fusion of global information with multiscale local features. Furthermore, we created a reverse dynamic feature fusion module (RDFM) at the decoding stage to efficiently fuse features at various levels while taking into account the integrity and specifics of the lesion region to increase the precision of complex lesion region segmentation. We carried out in-depth tests on three publicly accessible datasets International Skin Imaging Collaboration (ISIC)-2016, ISIC-2018, and PH2 to assess the method's efficacy and contrasted the outcomes with those of the most advanced techniques; the results confirmed that the suggested approach was superior.

为了诊断疾病和安排治疗,必须准确、一致地对皮肤镜图像中的皮肤病变进行分割。然而,在处理模糊的病灶区域边界、多尺度特征以及病灶区域的大小、形状和颜色变化明显等问题时,现有方法通常具有较高的计算复杂度和较大的参数计数。由于对局部特征和全局信息的捕捉不足,它们也经常遭受分割精度下降的困扰。为了克服这些问题,本文提出了一种基于高性能自适应注意力的轻量级深度学习网络。具体来说,引入了深度卷积神经网络来捕获局部信息。同时,构建了高性能的自适应关注特征融合模块(EAAF),利用动态特征选择实现全局信息与多尺度局部特征的自适应融合。此外,我们在解码阶段创建了反向动态特征融合模块(RDFM),在考虑病变区域的完整性和特殊性的同时,有效地融合了各级特征,提高了复杂病变区域分割的精度。我们对国际皮肤成像合作组织(ISIC)-2016、ISIC-2018和PH2三个可公开访问的数据集进行了深入测试,以评估该方法的疗效,并将结果与最先进的技术进行了对比;结果证实了该方法的优越性。
{"title":"HAI-Net: Skin Lesion Segmentation Using a High-Performance Adaptive Attention and Information Interaction Network","authors":"Chao Fan,&nbsp;Li Chen,&nbsp;Mengyang Yun,&nbsp;Huijun Zhao,&nbsp;Bincheng Peng","doi":"10.1002/ima.70266","DOIUrl":"https://doi.org/10.1002/ima.70266","url":null,"abstract":"<div>\u0000 \u0000 <p>Skin lesion segmentation from dermoscopic images must be done accurately and consistently in order to diagnose diseases and arrange treatments. However, when dealing with issues like fuzzy lesion region boundaries, multiscale features, and notable variations in the lesion region's size, shape, and color, existing methods typically have high computational complexity and large parameter counts. They also frequently suffer from decreased segmentation accuracy due to inadequate capture of local features and global information. In this paper, a lightweight deep learning network based on high-performance adaptive attention is proposed to overcome these issues. Specifically, a deep convolutional neural network is introduced to capture local information. Meanwhile, we create a high-performance adaptive attention feature fusion module (EAAF) that uses dynamic feature selection to achieve adaptive fusion of global information with multiscale local features. Furthermore, we created a reverse dynamic feature fusion module (RDFM) at the decoding stage to efficiently fuse features at various levels while taking into account the integrity and specifics of the lesion region to increase the precision of complex lesion region segmentation. We carried out in-depth tests on three publicly accessible datasets International Skin Imaging Collaboration (ISIC)-2016, ISIC-2018, and PH<sup>2</sup> to assess the method's efficacy and contrasted the outcomes with those of the most advanced techniques; the results confirmed that the suggested approach was superior.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750881","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SEA-Net: Dual Attention U-Net for Bleeding Segmentation in Capsule Endoscopy Images SEA-Net:用于胶囊内窥镜图像出血分割的双注意力U-Net
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-10 DOI: 10.1002/ima.70261
Tareque Bashar Ovi, Nomaiya Bashree, Hussain Nyeem, Md Abdul Wahed, Faiaz Hasanuzzaman Rhythm, Disha Chowdhury

Gastrointestinal (GI) bleeding, arising from various conditions, can be critical if untreated. Wireless capsule endoscopy (WCE) is a highly effective method for detecting GI bleeding, offering full visualization of the GI tract. However, the large number of images generated per patient poses challenges for clinicians, leading to prolonged analysis times and increased risk of human error. This emphasizes the need for computer-aided diagnosis systems. In this study, we introduce SEA-Net (Structured Efficient Attention Network), a novel deep learning network for detecting bleeding regions in WCE images. SEA-Net integrates a Convolutional Block Attention Module (CBAM) with long skip connections to enhance gradient flow and improve blood region localization. The EfficientNet-B4 encoder improves feature extraction efficiency and generalizability. A five-fold cross validation demonstrates consistent performance, while generalization tests, including precision-recall curves, ROC curves, and F1 measure, further validate the model's robustness. Minimal performance degradation was observed when the training data was reduced from 80% to 20%. Experimental results show that SEA-Net achieves a Dice score of 93.64% and an IoU score of 88.61% on a publicly available WCE dataset, outperforming state-of-the-art models and highlighting its strong potential for clinical application.

胃肠道(GI)出血,由各种情况引起,如果不治疗,可能会很严重。无线胶囊内窥镜(WCE)是一种非常有效的检测胃肠道出血的方法,提供了胃肠道的全面可视化。然而,每位患者生成的大量图像给临床医生带来了挑战,导致分析时间延长,人为错误的风险增加。这强调了对计算机辅助诊断系统的需求。在本研究中,我们引入了一种新的深度学习网络SEA-Net (Structured Efficient Attention Network),用于检测WCE图像中的出血区域。SEA-Net集成了一个带有长跳跃连接的卷积块注意模块(CBAM),以增强梯度流动和改善血液区域定位。高效网- b4编码器提高了特征提取效率和通用性。五重交叉验证显示了一致的性能,而包括精度-召回率曲线、ROC曲线和F1测量在内的泛化测试进一步验证了模型的稳健性。当训练数据从80%减少到20%时,观察到的性能下降最小。实验结果表明,SEA-Net在公开可用的WCE数据集上的Dice得分为93.64%,IoU得分为88.61%,优于目前最先进的模型,显示了其强大的临床应用潜力。
{"title":"SEA-Net: Dual Attention U-Net for Bleeding Segmentation in Capsule Endoscopy Images","authors":"Tareque Bashar Ovi,&nbsp;Nomaiya Bashree,&nbsp;Hussain Nyeem,&nbsp;Md Abdul Wahed,&nbsp;Faiaz Hasanuzzaman Rhythm,&nbsp;Disha Chowdhury","doi":"10.1002/ima.70261","DOIUrl":"https://doi.org/10.1002/ima.70261","url":null,"abstract":"<div>\u0000 \u0000 <p>Gastrointestinal (GI) bleeding, arising from various conditions, can be critical if untreated. Wireless capsule endoscopy (WCE) is a highly effective method for detecting GI bleeding, offering full visualization of the GI tract. However, the large number of images generated per patient poses challenges for clinicians, leading to prolonged analysis times and increased risk of human error. This emphasizes the need for computer-aided diagnosis systems. In this study, we introduce SEA-Net (<b>S</b>tructured <b>E</b>fficient <b>A</b>ttention <b>Net</b>work), a novel deep learning network for detecting bleeding regions in WCE images. SEA-Net integrates a Convolutional Block Attention Module (CBAM) with long skip connections to enhance gradient flow and improve blood region localization. The EfficientNet-B4 encoder improves feature extraction efficiency and generalizability. A five-fold cross validation demonstrates consistent performance, while generalization tests, including precision-recall curves, ROC curves, and F1 measure, further validate the model's robustness. Minimal performance degradation was observed when the training data was reduced from 80% to 20%. Experimental results show that SEA-Net achieves a Dice score of 93.64% and an IoU score of 88.61% on a publicly available WCE dataset, outperforming state-of-the-art models and highlighting its strong potential for clinical application.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Deep Fuzzy Inference System for Interpretable Multi-Class Heart Disease Risk Prediction 可解释多类别心脏病风险预测的深度模糊推理系统
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-10 DOI: 10.1002/ima.70264
S. Ramasami, P. Uma Maheswari

Heart disease remains a leading global health concern, necessitating accurate and interpretable risk prediction models for effective clinical decision-making. Accurate heart disease risk prediction is crucial for preventive healthcare, yet traditional machine learning models often struggle with the inherent uncertainty and nonlinear patterns in medical data. While deep neural networks (DNNs) excel at feature extraction, their lack of interpretability limits clinical utility, whereas fuzzy inference systems (FIS) offer transparency but lack hierarchical learning capabilities. To bridge this gap, we propose a novel Deep Fuzzy Inference System (DFIS) that integrates DNNs and FIS into a unified architecture, combining the strengths of both approaches. The DFIS leverages a weighted fusion mechanism to combine probabilistic outputs from an Adam Cuckoo Search-optimized DNN and a trapezoidal membership-based FIS, enabling simultaneous high accuracy and interpretability. The performance is evaluated on the Cleveland heart disease dataset; the DFIS achieves 97.2% accuracy, outperforming a standalone DNN (95.4%) and ANFIS (91.7%) under the same experimental conditions while providing clinically actionable risk stratification into normal, less critical, and very critical categories.

心脏病仍然是全球主要的健康问题,需要准确和可解释的风险预测模型来进行有效的临床决策。准确的心脏病风险预测对于预防性医疗保健至关重要,但传统的机器学习模型经常与医疗数据中固有的不确定性和非线性模式作斗争。虽然深度神经网络(dnn)擅长特征提取,但其缺乏可解释性限制了临床应用,而模糊推理系统(FIS)提供透明度,但缺乏分层学习能力。为了弥补这一差距,我们提出了一种新的深度模糊推理系统(DFIS),它将dnn和FIS集成到一个统一的架构中,结合了两种方法的优势。DFIS利用加权融合机制,将亚当布谷鸟搜索优化的深度神经网络和基于梯形隶属度的FIS的概率输出结合起来,同时实现高精度和可解释性。在克利夫兰心脏病数据集上对性能进行评估;在相同的实验条件下,DFIS达到97.2%的准确率,优于独立DNN(95.4%)和ANFIS(91.7%),同时提供临床可操作的风险分层,分为正常、不太严重和非常严重的类别。
{"title":"A Deep Fuzzy Inference System for Interpretable Multi-Class Heart Disease Risk Prediction","authors":"S. Ramasami,&nbsp;P. Uma Maheswari","doi":"10.1002/ima.70264","DOIUrl":"https://doi.org/10.1002/ima.70264","url":null,"abstract":"<div>\u0000 \u0000 <p>Heart disease remains a leading global health concern, necessitating accurate and interpretable risk prediction models for effective clinical decision-making. Accurate heart disease risk prediction is crucial for preventive healthcare, yet traditional machine learning models often struggle with the inherent uncertainty and nonlinear patterns in medical data. While deep neural networks (DNNs) excel at feature extraction, their lack of interpretability limits clinical utility, whereas fuzzy inference systems (FIS) offer transparency but lack hierarchical learning capabilities. To bridge this gap, we propose a novel Deep Fuzzy Inference System (DFIS) that integrates DNNs and FIS into a unified architecture, combining the strengths of both approaches. The DFIS leverages a weighted fusion mechanism to combine probabilistic outputs from an Adam Cuckoo Search-optimized DNN and a trapezoidal membership-based FIS, enabling simultaneous high accuracy and interpretability. The performance is evaluated on the Cleveland heart disease dataset; the DFIS achieves 97.2% accuracy, outperforming a standalone DNN (95.4%) and ANFIS (91.7%) under the same experimental conditions while providing clinically actionable risk stratification into normal, less critical, and very critical categories.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Lung Disease Diagnosis: A High Performance Hybrid Deep Learning Framework for Multi-Class Chest X-Ray Analysis 增强肺部疾病诊断:用于多类别胸部x射线分析的高性能混合深度学习框架
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-09 DOI: 10.1002/ima.70262
Tolga Saim Bascetin, Ibrahim Emiroglu

This study presents a high performance hybrid deep learning model for the classification of 14 lung diseases using chest X-ray (CXR) images. Manual evaluation of CXR images is labor-intensive and prone to human error. Therefore, automated systems are required to improve diagnostic accuracy and efficiency. Our model integrates ResNet18 and EfficientNet-V2-S architectures, combining residual connections with efficient scaling to achieve high accuracy while maintaining computational efficiency. Trained on the NIH ChestX-ray14 dataset, comprising 112 120 images across 14 disease classes, the model mitigates class imbalances with extensive data augmentation techniques. Achieving an impressive average AUC of 0.872, the model outperforms previous approaches. This performance was enhanced by a refined, anatomically-aware data augmentation strategy that improved the model's robustness and clinical relevance, particularly in challenging disease categories such as Pneumothorax, Emphysema, and Hernia. To further validate its generalizability, the proposed model was tested on three additional datasets for pneumonia, COVID-19, and tuberculosis. The results demonstrate superior performance, achieving an accuracy of 0.958, F1 score of 0.944, and ROC AUC of 0.989 for pneumonia; an accuracy of 0.974, F1 score of 0.969, and ROC AUC of 0.995 for COVID-19; and an accuracy of 0.999, F1 score of 0.999, and ROC AUC of 0.999 for tuberculosis. These outstanding results confirm the robustness and clinical applicability of the model across diverse datasets. This research introduces a reliable and efficient diagnostic tool that enhances the potential of automated lung disease classification. By alleviating radiologists' workload and promoting timely, accurate diagnostic outcomes, the model contributes significantly to medical imaging applications and demonstrates its capacity for practical use in real-world clinical settings.

本研究提出了一种高性能混合深度学习模型,用于使用胸部x射线(CXR)图像对14种肺部疾病进行分类。人工评估CXR图像是一项劳动密集型工作,而且容易出现人为错误。因此,需要自动化系统来提高诊断的准确性和效率。我们的模型集成了ResNet18和EfficientNet-V2-S架构,将剩余连接与高效缩放相结合,在保持计算效率的同时实现高精度。该模型在NIH ChestX-ray14数据集上进行训练,该数据集包含14种疾病类别的112 - 120张图像,通过广泛的数据增强技术减轻了类别不平衡。该模型获得了令人印象深刻的平均AUC 0.872,优于以前的方法。这一性能通过一种精细的、具有解剖学意识的数据增强策略得到了增强,该策略提高了模型的稳健性和临床相关性,特别是在气胸、肺气肿和疝气等具有挑战性的疾病类别中。为了进一步验证其普遍性,在肺炎、COVID-19和结核病的另外三个数据集上测试了所提出的模型。结果显示,该方法对肺炎的诊断准确率为0.958,F1评分为0.944,ROC AUC为0.989;COVID-19的准确率为0.974,F1评分为0.969,ROC AUC为0.995;诊断肺结核的准确率为0.999,F1评分为0.999,ROC AUC为0.999。这些突出的结果证实了该模型在不同数据集上的稳健性和临床适用性。本研究介绍了一种可靠、高效的诊断工具,提高了肺部疾病自动分类的潜力。通过减轻放射科医生的工作量,促进及时、准确的诊断结果,该模型为医学成像应用做出了重大贡献,并展示了其在现实世界临床环境中的实际应用能力。
{"title":"Enhancing Lung Disease Diagnosis: A High Performance Hybrid Deep Learning Framework for Multi-Class Chest X-Ray Analysis","authors":"Tolga Saim Bascetin,&nbsp;Ibrahim Emiroglu","doi":"10.1002/ima.70262","DOIUrl":"https://doi.org/10.1002/ima.70262","url":null,"abstract":"<div>\u0000 \u0000 <p>This study presents a high performance hybrid deep learning model for the classification of 14 lung diseases using chest X-ray (CXR) images. Manual evaluation of CXR images is labor-intensive and prone to human error. Therefore, automated systems are required to improve diagnostic accuracy and efficiency. Our model integrates ResNet18 and EfficientNet-V2-S architectures, combining residual connections with efficient scaling to achieve high accuracy while maintaining computational efficiency. Trained on the NIH ChestX-ray14 dataset, comprising 112 120 images across 14 disease classes, the model mitigates class imbalances with extensive data augmentation techniques. Achieving an impressive average AUC of 0.872, the model outperforms previous approaches. This performance was enhanced by a refined, anatomically-aware data augmentation strategy that improved the model's robustness and clinical relevance, particularly in challenging disease categories such as Pneumothorax, Emphysema, and Hernia. To further validate its generalizability, the proposed model was tested on three additional datasets for pneumonia, COVID-19, and tuberculosis. The results demonstrate superior performance, achieving an accuracy of 0.958, <i>F</i>1 score of 0.944, and ROC AUC of 0.989 for pneumonia; an accuracy of 0.974, <i>F</i>1 score of 0.969, and ROC AUC of 0.995 for COVID-19; and an accuracy of 0.999, <i>F</i>1 score of 0.999, and ROC AUC of 0.999 for tuberculosis. These outstanding results confirm the robustness and clinical applicability of the model across diverse datasets. This research introduces a reliable and efficient diagnostic tool that enhances the potential of automated lung disease classification. By alleviating radiologists' workload and promoting timely, accurate diagnostic outcomes, the model contributes significantly to medical imaging applications and demonstrates its capacity for practical use in real-world clinical settings.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Evidential Fusion of Light–Dark Features With Multi-Scan Mamba for Automated Macular Edema Diagnosis 自适应证据融合的光-暗特征与多扫描曼巴自动诊断黄斑水肿
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-09 DOI: 10.1002/ima.70265
Yiming Zhang, Hongqing Zhu, Tianwei Qian, Tong Hou, Ning Chen, Xun Xu, Bingcang Huang

Macular edema is a retinal disorder that can lead to significant vision loss, underscoring the need for accurate and intelligent automated diagnosis. However, its subtle manifestations in color fundus photography (CFP) pose considerable challenges for conventional deep learning models. In this work, we propose a novel diagnostic framework that integrates Dempster–Shafer (D–S) evidence theory—a principled approach for uncertainty quantification and multi-source information fusion—with the advanced Mamba architecture. The proposed method employs a dual-branch network to selectively enhance and extract discriminative features from both bright and dark regions of fundus images. These features are dynamically aligned and fused via an Adaptive Multi-Branch Feature Synthesis (AMFS) module. To model long-range dependencies and aggregate complementary information from multiple scanning views, we introduce a multi-scan Mamba module, whose outputs are further fused using a principled D–S evidence mechanism. This synergistic integration of mathematical theory and deep learning not only reduces information redundancy but also enables confidence-aware automated decision-making. Extensive experiments on three retinal image datasets—IDRiD, Messidor, and a proprietary clinical cohort—demonstrate that our framework consistently outperforms state-of-the-art methods in terms of accuracy, F1-score, and robustness. These results highlight the promise of combining evidence theory with modern deep learning for challenging medical image analysis tasks.

黄斑水肿是一种视网膜疾病,可导致严重的视力丧失,强调需要准确和智能的自动诊断。然而,它在彩色眼底摄影(CFP)中的细微表现对传统的深度学习模型提出了相当大的挑战。在这项工作中,我们提出了一个新的诊断框架,该框架将Dempster-Shafer (D-S)证据理论(一种用于不确定性量化和多源信息融合的原则性方法)与先进的Mamba架构集成在一起。该方法采用双分支网络对眼底图像的明暗区域进行选择性增强和特征提取。这些特征通过自适应多分支特征合成(AMFS)模块动态对齐和融合。为了对远程依赖关系进行建模,并从多个扫描视图中汇总互补信息,我们引入了一个多扫描Mamba模块,其输出使用原则性D-S证据机制进一步融合。这种数学理论和深度学习的协同集成不仅减少了信息冗余,而且还实现了自信感知的自动决策。在三个视网膜图像数据集(idrid、Messidor和一个专有的临床队列)上进行的广泛实验表明,我们的框架在准确性、f1评分和稳健性方面始终优于最先进的方法。这些结果突出了将证据理论与现代深度学习相结合以解决具有挑战性的医学图像分析任务的前景。
{"title":"Adaptive Evidential Fusion of Light–Dark Features With Multi-Scan Mamba for Automated Macular Edema Diagnosis","authors":"Yiming Zhang,&nbsp;Hongqing Zhu,&nbsp;Tianwei Qian,&nbsp;Tong Hou,&nbsp;Ning Chen,&nbsp;Xun Xu,&nbsp;Bingcang Huang","doi":"10.1002/ima.70265","DOIUrl":"https://doi.org/10.1002/ima.70265","url":null,"abstract":"<div>\u0000 \u0000 <p>Macular edema is a retinal disorder that can lead to significant vision loss, underscoring the need for accurate and intelligent automated diagnosis. However, its subtle manifestations in color fundus photography (CFP) pose considerable challenges for conventional deep learning models. In this work, we propose a novel diagnostic framework that integrates Dempster–Shafer (D–S) evidence theory—a principled approach for uncertainty quantification and multi-source information fusion—with the advanced Mamba architecture. The proposed method employs a dual-branch network to selectively enhance and extract discriminative features from both bright and dark regions of fundus images. These features are dynamically aligned and fused via an Adaptive Multi-Branch Feature Synthesis (AMFS) module. To model long-range dependencies and aggregate complementary information from multiple scanning views, we introduce a multi-scan Mamba module, whose outputs are further fused using a principled D–S evidence mechanism. This synergistic integration of mathematical theory and deep learning not only reduces information redundancy but also enables confidence-aware automated decision-making. Extensive experiments on three retinal image datasets—IDRiD, Messidor, and a proprietary clinical cohort—demonstrate that our framework consistently outperforms state-of-the-art methods in terms of accuracy, F1-score, and robustness. These results highlight the promise of combining evidence theory with modern deep learning for challenging medical image analysis tasks.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Task-Specific Electroencephlogram Analysis: A Novel ICA and Dynamic Multi-Stage Clustering Approach for Neural Signal Processing 任务特异性脑电图分析:一种新的ICA和动态多阶段聚类方法用于神经信号处理
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-29 DOI: 10.1002/ima.70260
C. Kaviyazhiny, P. Shanthi Bala, R. Priyadharshini, S. Ajeeth

Electroencephalogram (EEG) plays a vital role in identifying the neural activities and human behaviors. EEG is a high-temporal-resolution signal that is contaminated and influenced by various external factors and artifacts. It is very difficult to identify and analyze the EEG signal when it is contaminated by artifacts. The impact of artifacts in EEG signals became an open challenge in EEG signal processing. Even though many techniques and methods are available to remove the artifacts in EEG, analyzing the contaminated EEG signals remains a significant challenge. The signal is polluted by different artifacts like eye blink, muscle activities, environmental interfaces, and so on. Compared with other artifacts and noises, task-specific artifacts such as motor movement and motor imagery have a greater impact on EEG signals due to their direct inference on cognitive signals, and isolating task-specific artifacts from the EEG signal is difficult. To overcome these challenges, a novel hybrid task-specific Dynamic Multistage k-Means Clustering algorithm (TS-DMKC) has been proposed that detects and extracts the processed EEG signal, which will be helpful for all the EEG applications. Initially, the motor movement and motor imagery artifacts are distinguished using the Fast ICA by applying the filtering threshold. Then, a multistage k-means clustering algorithm is employed to isolate the artifacts in different clusters dynamically, and the cluster effectiveness is calculated by using the silhouette score. This proposed novel hybrid algorithm will be useful in various EEG applications like neuroscience, medical diagnosis, brain–computer interface, authentication, brain signal analysis, and gaming. The experimental results demonstrate that the proposed algorithm TS-DMKC achieved a classification accuracy of 94.2% using the Physionet motor movement/imagery dataset, reducing task-dependent variability and offering more robust and reliable EEG systems.

脑电图(EEG)在识别神经活动和人类行为方面起着至关重要的作用。脑电图是一种高时间分辨率的信号,受到各种外界因素和人为因素的污染和影响。当脑电信号受到人为干扰时,对其进行识别和分析是非常困难的。脑电信号中伪影的影响成为脑电信号处理中的一个开放性难题。尽管有许多技术和方法可以去除脑电信号中的伪影,但分析受污染的脑电信号仍然是一个重大挑战。信号会受到不同的干扰,比如眨眼、肌肉活动、环境界面等等。与其他伪像和噪声相比,特定任务伪像(如运动运动和运动意象)对脑电信号的影响更大,因为它们直接对认知信号进行推断,并且很难从脑电信号中分离出特定任务伪像。为了克服这些挑战,本文提出了一种新的针对特定任务的混合动态多阶段k-均值聚类算法(TS-DMKC)来检测和提取处理后的脑电信号,这将有助于脑电信号的各种应用。首先,使用Fast ICA通过应用滤波阈值来区分运动运动和运动图像伪影。然后,采用多阶段k-means聚类算法动态分离不同聚类中的伪影,并利用轮廓分数计算聚类有效性;该算法可应用于神经科学、医学诊断、脑机接口、身份验证、脑信号分析、游戏等领域。实验结果表明,所提出的算法TS-DMKC在使用Physionet运动/图像数据集的情况下,达到了94.2%的分类准确率,减少了任务相关的可变性,提供了更加鲁棒和可靠的脑电系统。
{"title":"Task-Specific Electroencephlogram Analysis: A Novel ICA and Dynamic Multi-Stage Clustering Approach for Neural Signal Processing","authors":"C. Kaviyazhiny,&nbsp;P. Shanthi Bala,&nbsp;R. Priyadharshini,&nbsp;S. Ajeeth","doi":"10.1002/ima.70260","DOIUrl":"https://doi.org/10.1002/ima.70260","url":null,"abstract":"<div>\u0000 \u0000 <p>Electroencephalogram (EEG) plays a vital role in identifying the neural activities and human behaviors. EEG is a high-temporal-resolution signal that is contaminated and influenced by various external factors and artifacts. It is very difficult to identify and analyze the EEG signal when it is contaminated by artifacts. The impact of artifacts in EEG signals became an open challenge in EEG signal processing. Even though many techniques and methods are available to remove the artifacts in EEG, analyzing the contaminated EEG signals remains a significant challenge. The signal is polluted by different artifacts like eye blink, muscle activities, environmental interfaces, and so on. Compared with other artifacts and noises, task-specific artifacts such as motor movement and motor imagery have a greater impact on EEG signals due to their direct inference on cognitive signals, and isolating task-specific artifacts from the EEG signal is difficult. To overcome these challenges, a novel hybrid task-specific Dynamic Multistage <i>k</i>-Means Clustering algorithm (TS-DMKC) has been proposed that detects and extracts the processed EEG signal, which will be helpful for all the EEG applications. Initially, the motor movement and motor imagery artifacts are distinguished using the Fast ICA by applying the filtering threshold. Then, a multistage <i>k</i>-means clustering algorithm is employed to isolate the artifacts in different clusters dynamically, and the cluster effectiveness is calculated by using the silhouette score. This proposed novel hybrid algorithm will be useful in various EEG applications like neuroscience, medical diagnosis, brain–computer interface, authentication, brain signal analysis, and gaming. The experimental results demonstrate that the proposed algorithm TS-DMKC achieved a classification accuracy of 94.2% using the Physionet motor movement/imagery dataset, reducing task-dependent variability and offering more robust and reliable EEG systems.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145619201","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HPPE-Unet: A Boundary-Enhanced U-Shaped Network With Hybrid Attention and Multi-Scale Feature Fusion for Small-Sample Medical Image Segmentation hpe - unet:一种混合关注和多尺度特征融合的边界增强u型网络用于小样本医学图像分割
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-20 DOI: 10.1002/ima.70256
Changhao Sun, Tong Liu, Mujun Zang, Shusen Zhou, Chanjuan Liu, Qingjun Wang

Accurate diagnosis of diseases relies on the subjective experience of physicians. Automatic image segmentation technology can assist doctors in quickly locating lesion areas, thereby improving diagnostic efficiency. In the field of medical image segmentation, the current mainstream models are mostly variants of Swin-Unet, which perform well on large-sample datasets. However, due to the window constraints of the Swin-Transformer architecture, these models often face performance limitations when processing small-sample datasets. In contrast, convolutional-based Unet variant models exhibit better segmentation performance under small-sample conditions but demonstrate poorer segmentation results for medical images with blurred boundaries. To tackle these challenges, this paper introduces a novel boundary information-enhanced architecture, the Hybrid-Parallel-Attention and Pyramid-Edge-Extraction Enhanced UNet (HPPE-Unet). Based on an encoder-decoder architecture of convolutional neural networks, the model incorporates a multi-scale feature extraction strategy and an optimized skip-connection mechanism. It integrates two key modules: the Pyramid Edge Extraction (PEE) module and the Hybrid Parallel Attention (HP) module. The PEE module is applied at each stage of the encoder and decoder, significantly enhancing the model's ability to capture subtle boundary structures through multi-scale feature fusion and boundary reinforcement mechanisms. The HP employs a residual parallel structure combining spatial (SA) and channel attention (CA) blocks, effectively bridging the semantic gap between features in the encoding and decoding stages and addressing the issue of information loss in skip connections between the encoder and decoder. The proposed method was evaluated on an aortic dissection dataset provided by a tertiary hospital. The results show that the method achieved a Dice similarity coefficient (DSC) of 97.65% and a Mean intersection over union (Miou) of 97.92% in the segmentation task. On the public Automated Cardiac Diagnostic Challenge (ACDC) dataset, the DSC reached 90.39%. These results demonstrate that the proposed method holds significant practical value for clinical disease diagnosis.

疾病的准确诊断依赖于医师的主观经验。自动图像分割技术可以帮助医生快速定位病变区域,从而提高诊断效率。在医学图像分割领域,目前主流的模型大多是swan - unet的变体,在大样本数据集上表现良好。然而,由于swing - transformer架构的窗口约束,这些模型在处理小样本数据集时经常面临性能限制。相比之下,基于卷积的Unet变体模型在小样本条件下表现出更好的分割性能,但在边界模糊的医学图像中表现出较差的分割效果。为了解决这些问题,本文提出了一种新的边界信息增强架构——混合并行注意和金字塔边缘提取增强UNet (hpe - UNet)。该模型基于卷积神经网络的编码器-解码器架构,结合了多尺度特征提取策略和优化的跳跃连接机制。它集成了两个关键模块:金字塔边缘提取(PEE)模块和混合并行注意(HP)模块。在编码器和解码器的每个阶段都应用了PEE模块,通过多尺度特征融合和边界强化机制,显著增强了模型捕捉细微边界结构的能力。HP采用结合空间(SA)和信道注意(CA)块的残差并行结构,有效地弥合了编码和解码阶段特征之间的语义差距,并解决了编码器和解码器之间的跳过连接中的信息丢失问题。该方法在一家三级医院提供的主动脉夹层数据集上进行了评估。结果表明,该方法在分割任务中获得了97.65%的Dice similarity coefficient (DSC)和97.92%的Mean intersection over union (Miou)。在公共自动心脏诊断挑战(ACDC)数据集上,DSC达到90.39%。结果表明,该方法对临床疾病诊断具有重要的实用价值。
{"title":"HPPE-Unet: A Boundary-Enhanced U-Shaped Network With Hybrid Attention and Multi-Scale Feature Fusion for Small-Sample Medical Image Segmentation","authors":"Changhao Sun,&nbsp;Tong Liu,&nbsp;Mujun Zang,&nbsp;Shusen Zhou,&nbsp;Chanjuan Liu,&nbsp;Qingjun Wang","doi":"10.1002/ima.70256","DOIUrl":"https://doi.org/10.1002/ima.70256","url":null,"abstract":"<div>\u0000 \u0000 <p>Accurate diagnosis of diseases relies on the subjective experience of physicians. Automatic image segmentation technology can assist doctors in quickly locating lesion areas, thereby improving diagnostic efficiency. In the field of medical image segmentation, the current mainstream models are mostly variants of Swin-Unet, which perform well on large-sample datasets. However, due to the window constraints of the Swin-Transformer architecture, these models often face performance limitations when processing small-sample datasets. In contrast, convolutional-based Unet variant models exhibit better segmentation performance under small-sample conditions but demonstrate poorer segmentation results for medical images with blurred boundaries. To tackle these challenges, this paper introduces a novel boundary information-enhanced architecture, the Hybrid-Parallel-Attention and Pyramid-Edge-Extraction Enhanced UNet (HPPE-Unet). Based on an encoder-decoder architecture of convolutional neural networks, the model incorporates a multi-scale feature extraction strategy and an optimized skip-connection mechanism. It integrates two key modules: the Pyramid Edge Extraction (PEE) module and the Hybrid Parallel Attention (HP) module. The PEE module is applied at each stage of the encoder and decoder, significantly enhancing the model's ability to capture subtle boundary structures through multi-scale feature fusion and boundary reinforcement mechanisms. The HP employs a residual parallel structure combining spatial (SA) and channel attention (CA) blocks, effectively bridging the semantic gap between features in the encoding and decoding stages and addressing the issue of information loss in skip connections between the encoder and decoder. The proposed method was evaluated on an aortic dissection dataset provided by a tertiary hospital. The results show that the method achieved a Dice similarity coefficient (DSC) of 97.65% and a Mean intersection over union (Miou) of 97.92% in the segmentation task. On the public Automated Cardiac Diagnostic Challenge (ACDC) dataset, the DSC reached 90.39%. These results demonstrate that the proposed method holds significant practical value for clinical disease diagnosis.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145572471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncertainty-Aware Graph Self-Training for Autism Spectrum Disorder Classification in Multiple Centers 自闭症谱系障碍多中心分类的不确定性感知图自训练
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-20 DOI: 10.1002/ima.70250
Qianhui Yang, Jun Wang, Jiale Dun, Juncheng Li, Jun Shi

Classical self-training methods for graph convolutional networks (GCNs) assume that both labeled and unlabeled data follow the identical distribution. However, these works do not work well when they are used in medical applications such as classifying autism spectrum disorder (ASD) in multiple centers, where the unlabeled samples from different imaging centers have varying distribution shifts from the labeled samples. To this end, we propose uncertainty-aware graph self-training (UA-GST) by extending graph self-training to a new situation, in which the labeled data come from one imaging center and the unlabeled data come from several other imaging centers. Specifically, an uncertainty-aware mechanism is proposed to select unlabeled centers and adversarial domain adaptation is introduced into graph self-training to reduce domain shift between centers. With the progression of self-training, more pseudo-labeled test samples are gradually included in the training set, and a final model is finally trained on all the labeled and pseudo-labeled samples. Considering the over-confidence issue of the classifier, an evidential GCN is further proposed to estimate the uncertainty of the pseudo-labels using Dempster–Shafer (D–S) evidence theory. It is evaluated on the Autism Brain Imaging Data Exchange (ABIDE). Experimental results verified the effectiveness of the proposed method in classifying ASD in multiple centers, outperforming existing state-of-the-art methods.

经典的图卷积网络(GCNs)自训练方法假设标记数据和未标记数据遵循相同的分布。然而,当这些工作被用于医学应用时,如在多个中心对自闭症谱系障碍(ASD)进行分类时,这些工作就不能很好地发挥作用,因为来自不同成像中心的未标记样本与标记样本的分布变化不同。为此,我们提出了不确定性感知图自训练(UA-GST),将图自训练扩展到一种新的情况,即标记数据来自一个成像中心,未标记数据来自其他几个成像中心。具体而言,提出了一种不确定性感知机制来选择未标记的中心,并在图自训练中引入对抗域适应来减少中心之间的域移动。随着自我训练的进行,越来越多的伪标记测试样本逐渐被纳入训练集,最后在所有标记和伪标记样本上训练出最终的模型。考虑到分类器的过度置信度问题,利用Dempster-Shafer (D-S)证据理论,进一步提出了一种估计伪标签不确定性的证据GCN。在自闭症脑成像数据交换(Autism Brain Imaging Data Exchange,简称ABIDE)上进行评估。实验结果验证了该方法在多个中心的ASD分类中的有效性,优于现有的最先进的方法。
{"title":"Uncertainty-Aware Graph Self-Training for Autism Spectrum Disorder Classification in Multiple Centers","authors":"Qianhui Yang,&nbsp;Jun Wang,&nbsp;Jiale Dun,&nbsp;Juncheng Li,&nbsp;Jun Shi","doi":"10.1002/ima.70250","DOIUrl":"https://doi.org/10.1002/ima.70250","url":null,"abstract":"<div>\u0000 \u0000 <p>Classical self-training methods for graph convolutional networks (GCNs) assume that both labeled and unlabeled data follow the identical distribution. However, these works do not work well when they are used in medical applications such as classifying autism spectrum disorder (ASD) in multiple centers, where the unlabeled samples from different imaging centers have varying distribution shifts from the labeled samples. To this end, we propose uncertainty-aware graph self-training (UA-GST) by extending graph self-training to a new situation, in which the labeled data come from one imaging center and the unlabeled data come from several other imaging centers. Specifically, an uncertainty-aware mechanism is proposed to select unlabeled centers and adversarial domain adaptation is introduced into graph self-training to reduce domain shift between centers. With the progression of self-training, more pseudo-labeled test samples are gradually included in the training set, and a final model is finally trained on all the labeled and pseudo-labeled samples. Considering the over-confidence issue of the classifier, an evidential GCN is further proposed to estimate the uncertainty of the pseudo-labels using Dempster–Shafer (D–S) evidence theory. It is evaluated on the Autism Brain Imaging Data Exchange (ABIDE). Experimental results verified the effectiveness of the proposed method in classifying ASD in multiple centers, outperforming existing state-of-the-art methods.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145572607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Supervised DRL With Twin Critics: A Novel Framework for Glioblastoma Survival Prognostication 双重批评的自我监督DRL:胶质母细胞瘤生存预测的新框架
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-20 DOI: 10.1002/ima.70258
M. Renugadevi, K. Ramkumar, N. Raju, K. Adalarasu, Surya Prasath, K. Narasimhan

The accurate survival prediction in glioblastoma patients remains a major challenge, largely owing to the considerable variability among brain tumor subtypes and their clinical profiles. This research introduces an innovative SimCLR-Twin Critic DDPG framework that integrates self-supervised representation learning with deep reinforcement learning for enhanced prognostic prediction. Tumor subregion segmentations were independently obtained using four advanced models namely nnU-Net, VNet, SwinUNETR, and UNet. Based on the performance of these segmentation models, deep features were extracted from the nnU-Net segmented tumor subregions and combined with handcrafted radiomics features. The combined features were subjected to the feature selection process using BorutaShap, followed by self-supervised pretraining through SimCLR to improve feature representation and generalization. The optimized features were then fed into a Twin Critic DDPG agent designed for regression-based survival time predictions. The proposed method outperformed existing techniques on the BraTS 2020 dataset, achieving the lowest MSE of 31 062.36 and the highest C-index of 0.87 in overall survival prediction. To further confirm robustness and generalizability, the framework was externally validated on the BraTS 2019 dataset, where it achieved a comparable MSE of 31 156.56 and a C-index of 0.86 using fused features. Additionally, LIME-based local interpretability provided clinically relevant explanations for individual predictions, thereby enhancing trust in the AI-driven system. This study highlights the SimCLR-Twin Critic DDPG framework as a robust and interpretable solution for accurate survival prediction in glioblastoma patients.

胶质母细胞瘤患者的准确生存预测仍然是一个主要的挑战,主要是由于脑肿瘤亚型及其临床特征的相当大的变异性。本研究引入了一个创新的SimCLR-Twin Critic DDPG框架,该框架集成了自监督表示学习和深度强化学习,以增强预后预测。采用nnU-Net、VNet、SwinUNETR和UNet四种高级模型独立获得肿瘤亚区分割。基于这些分割模型的性能,从nnU-Net分割的肿瘤亚区域中提取深度特征,并与手工制作的放射组学特征相结合。组合后的特征采用BorutaShap进行特征选择,然后采用SimCLR进行自监督预训练,提高特征表征和泛化能力。然后将优化的特征输入Twin Critic DDPG代理,该代理设计用于基于回归的生存时间预测。该方法在BraTS 2020数据集上优于现有技术,在总体生存预测中实现了最低的MSE为31 062.36,最高的c指数为0.87。为了进一步确认稳健性和泛化性,该框架在BraTS 2019数据集上进行了外部验证,其中使用融合特征实现了可比较的MSE为31 156.56和c指数为0.86。此外,基于lime的局部可解释性为个体预测提供了临床相关的解释,从而增强了对人工智能驱动系统的信任。该研究强调了SimCLR-Twin Critic DDPG框架作为胶质母细胞瘤患者准确生存预测的可靠且可解释的解决方案。
{"title":"Self-Supervised DRL With Twin Critics: A Novel Framework for Glioblastoma Survival Prognostication","authors":"M. Renugadevi,&nbsp;K. Ramkumar,&nbsp;N. Raju,&nbsp;K. Adalarasu,&nbsp;Surya Prasath,&nbsp;K. Narasimhan","doi":"10.1002/ima.70258","DOIUrl":"https://doi.org/10.1002/ima.70258","url":null,"abstract":"<div>\u0000 \u0000 <p>The accurate survival prediction in glioblastoma patients remains a major challenge, largely owing to the considerable variability among brain tumor subtypes and their clinical profiles. This research introduces an innovative SimCLR-Twin Critic DDPG framework that integrates self-supervised representation learning with deep reinforcement learning for enhanced prognostic prediction. Tumor subregion segmentations were independently obtained using four advanced models namely nnU-Net, VNet, SwinUNETR, and UNet. Based on the performance of these segmentation models, deep features were extracted from the nnU-Net segmented tumor subregions and combined with handcrafted radiomics features. The combined features were subjected to the feature selection process using BorutaShap, followed by self-supervised pretraining through SimCLR to improve feature representation and generalization. The optimized features were then fed into a Twin Critic DDPG agent designed for regression-based survival time predictions. The proposed method outperformed existing techniques on the BraTS 2020 dataset, achieving the lowest MSE of 31 062.36 and the highest C-index of 0.87 in overall survival prediction. To further confirm robustness and generalizability, the framework was externally validated on the BraTS 2019 dataset, where it achieved a comparable MSE of 31 156.56 and a C-index of 0.86 using fused features. Additionally, LIME-based local interpretability provided clinically relevant explanations for individual predictions, thereby enhancing trust in the AI-driven system. This study highlights the SimCLR-Twin Critic DDPG framework as a robust and interpretable solution for accurate survival prediction in glioblastoma patients.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145581098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Objectives Optimization (MOO)-Based NAS Approach 基于多目标优化(MOO)的NAS方法
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-15 DOI: 10.1002/ima.70240
Abass Sana, Kaoutar Senhaji, Amir Nakib

This research dives into creating optimized neural network architectures tailored for image classification and medical image segmentation, with a particular emphasis on analyzing lung nodules using the LIDC-IDRI dataset. We introduce a hybrid approach to neural architecture search (NAS) that fuses convolutional neural networks (CNNs), transformer-based models, and custom-built graph architectures. Our evolutionary strategies employ genetic operations like crossover and mutation to gradually enhance these architectures, while the NSGA-III algorithm helps us navigate the tricky balance of multiple conflicting objectives. The goal of our method is to boost performance metrics such as the F1 score and classification accuracy, all while keeping an eye on minimizing computational complexity, which we measure in terms of FLOPS, parameter count, and inference time. Our experiments demonstrate competitive performance on the CIFAR-10 and CIFAR-100 datasets, with promising results on the LIDC-IDRI segmentation task. This work aims to push the boundaries of automated model design, contributing to more efficient and effective deep learning architectures in both general and medical imaging contexts.

本研究深入研究了为图像分类和医学图像分割量身定制的优化神经网络架构,特别强调了使用LIDC-IDRI数据集分析肺结节。我们介绍了一种神经架构搜索(NAS)的混合方法,它融合了卷积神经网络(cnn)、基于变压器的模型和定制的图架构。我们的进化策略采用交叉和突变等遗传操作来逐步增强这些结构,而NSGA-III算法则帮助我们在多个相互冲突的目标之间找到棘手的平衡。我们方法的目标是提高F1分数和分类准确性等性能指标,同时关注最小化计算复杂性,我们用FLOPS、参数计数和推理时间来衡量计算复杂性。我们的实验在CIFAR-10和CIFAR-100数据集上展示了具有竞争力的性能,在LIDC-IDRI分割任务上取得了令人鼓舞的结果。这项工作旨在推动自动化模型设计的边界,为通用和医学成像环境中更高效和有效的深度学习架构做出贡献。
{"title":"Multi-Objectives Optimization (MOO)-Based NAS Approach","authors":"Abass Sana,&nbsp;Kaoutar Senhaji,&nbsp;Amir Nakib","doi":"10.1002/ima.70240","DOIUrl":"https://doi.org/10.1002/ima.70240","url":null,"abstract":"<div>\u0000 \u0000 <p>This research dives into creating optimized neural network architectures tailored for image classification and medical image segmentation, with a particular emphasis on analyzing lung nodules using the LIDC-IDRI dataset. We introduce a hybrid approach to neural architecture search (NAS) that fuses convolutional neural networks (CNNs), transformer-based models, and custom-built graph architectures. Our evolutionary strategies employ genetic operations like crossover and mutation to gradually enhance these architectures, while the NSGA-III algorithm helps us navigate the tricky balance of multiple conflicting objectives. The goal of our method is to boost performance metrics such as the F1 score and classification accuracy, all while keeping an eye on minimizing computational complexity, which we measure in terms of FLOPS, parameter count, and inference time. Our experiments demonstrate competitive performance on the CIFAR-10 and CIFAR-100 datasets, with promising results on the LIDC-IDRI segmentation task. This work aims to push the boundaries of automated model design, contributing to more efficient and effective deep learning architectures in both general and medical imaging contexts.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145530079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1