首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
EGCF-Net Edge-Aware Graph-Based Attention Network With Enhanced Contextual Features for Medical Image Segmentation 基于增强上下文特征的EGCF-Net边缘感知图关注网络用于医学图像分割
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-28 DOI: 10.1002/ima.70276
Santoshi Gorli, Ratnakar Dash

Accurate segmentation of medical images is critical for disease diagnosis and treatment planning. However, challenges such as fuzzy boundaries, low contrast, and complex anatomical structures often hinder performance. We propose EGCF-Net, a novel U-Net-based architecture that integrates a hybrid encoder with an edge-aware graph-based attention network to address these limitations. The hybrid encoder integrates the Swin transformer and Kronecker convolution to capture global and local contextual dependencies. Additionally, skip connections are enhanced using an edge-aware graph-based attention module, which combines graph spatial attention and graph channel attention to dynamically model spatial correlations and edge-aware contextual affinities. This design leads to edge-enhanced boundary delineation and improved regional consistency. We evaluate EGCF-Net on four benchmark datasets (Synapse, Kidney Stone, ISIC 2016, and ISIC 2018), achieving Dice scores of 84.70%, 93.07%, 91.07%, and 88.62%, respectively, surpassing existing state-of-the-art methods. Quantitative and qualitative results further validate the efficacy and robustness of the proposed approach, highlighting its potential for advancing medical image segmentation.

医学图像的准确分割是疾病诊断和治疗计划的关键。然而,诸如模糊边界、低对比度和复杂解剖结构等挑战往往会阻碍性能。我们提出了EGCF-Net,这是一种新颖的基于u - net的架构,它将混合编码器与基于边缘感知图的注意力网络集成在一起,以解决这些限制。混合编码器集成了Swin变压器和Kronecker卷积,以捕获全局和局部上下文依赖关系。此外,使用基于边缘感知的图形注意模块增强了跳过连接,该模块结合了图形空间注意和图形通道注意来动态建模空间相关性和边缘感知上下文亲和力。这种设计导致边缘增强边界划定和提高区域一致性。我们在四个基准数据集(Synapse, Kidney Stone, ISIC 2016和ISIC 2018)上对EGCF-Net进行了评估,分别获得了84.70%,93.07%,91.07%和88.62%的Dice分数,超过了现有的最先进的方法。定量和定性结果进一步验证了所提出方法的有效性和鲁棒性,突出了其在推进医学图像分割方面的潜力。
{"title":"EGCF-Net Edge-Aware Graph-Based Attention Network With Enhanced Contextual Features for Medical Image Segmentation","authors":"Santoshi Gorli,&nbsp;Ratnakar Dash","doi":"10.1002/ima.70276","DOIUrl":"https://doi.org/10.1002/ima.70276","url":null,"abstract":"<div>\u0000 \u0000 <p>Accurate segmentation of medical images is critical for disease diagnosis and treatment planning. However, challenges such as fuzzy boundaries, low contrast, and complex anatomical structures often hinder performance. We propose EGCF-Net, a novel U-Net-based architecture that integrates a hybrid encoder with an edge-aware graph-based attention network to address these limitations. The hybrid encoder integrates the Swin transformer and Kronecker convolution to capture global and local contextual dependencies. Additionally, skip connections are enhanced using an edge-aware graph-based attention module, which combines graph spatial attention and graph channel attention to dynamically model spatial correlations and edge-aware contextual affinities. This design leads to edge-enhanced boundary delineation and improved regional consistency. We evaluate EGCF-Net on four benchmark datasets (Synapse, Kidney Stone, ISIC 2016, and ISIC 2018), achieving Dice scores of 84.70%, 93.07%, 91.07%, and 88.62%, respectively, surpassing existing state-of-the-art methods. Quantitative and qualitative results further validate the efficacy and robustness of the proposed approach, highlighting its potential for advancing medical image segmentation.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145904991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
StackDeVNet: An Explainable Stacking Ensemble of DenseNets and Vision Transformers for Advanced Gastrointestinal Disease Detection 用于高级胃肠道疾病检测的可解释的densenet和视觉变压器堆叠集成
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-26 DOI: 10.1002/ima.70275
Osman Güler

Gastrointestinal disorders include diseases that negatively affect people's daily life and carry the risk of cancer. Therefore, accurate and early diagnosis of these diseases is important for treatment process of patients. Deep learning architectures, which have achieved significant success in medical image analysis, are effectively used in early diagnosis and diagnosis systems. Therefore, in this study, a new approach that achieves higher accuracy in the detection of gastrointestinal diseases by combining DenseNet and Vision Transformer models with stacking ensemble is proposed. As a result of the experiments, the proposed model achieved 99.06% accuracy in a single test and 98.64% accuracy on mean as a result of 5-fold cross-validation. The proposed approach shows promising accuracy and reliability as evidenced by the results of experiments on the KvasirV2 dataset, and has the potential to be an effective method for the detection of gastrointestinal diseases. To improve model interpretability, the Explainable AI technique Grad-CAM and attention map visualizations were used, allowing visual justification of the model's predictions and highlighting clinically relevant regions in endoscopic images. The model obtained by combining DenseNet and Vision Transformer models with the stacking ensemble method is expected to be an example for future studies in the field of health and image processing, especially gastrointestinal diseases.

胃肠道疾病包括对人们日常生活产生负面影响并具有癌症风险的疾病。因此,准确、早期诊断这些疾病对患者的治疗过程至关重要。深度学习架构在医学图像分析方面取得了显著的成功,可以有效地应用于早期诊断和诊断系统。因此,本研究提出了一种将DenseNet和Vision Transformer模型与堆叠集成相结合的方法,以达到更高的胃肠道疾病检测精度。实验结果表明,该模型在单次测试中准确率达到99.06%,经过5次交叉验证,平均准确率达到98.64%。KvasirV2数据集上的实验结果表明,该方法具有良好的准确性和可靠性,有可能成为检测胃肠道疾病的有效方法。为了提高模型的可解释性,使用了Explainable AI技术Grad-CAM和注意图可视化,允许对模型的预测进行视觉证明,并在内镜图像中突出显示临床相关区域。将DenseNet和Vision Transformer模型与叠加集成方法相结合得到的模型有望成为未来健康和图像处理领域,特别是胃肠道疾病领域研究的范例。
{"title":"StackDeVNet: An Explainable Stacking Ensemble of DenseNets and Vision Transformers for Advanced Gastrointestinal Disease Detection","authors":"Osman Güler","doi":"10.1002/ima.70275","DOIUrl":"https://doi.org/10.1002/ima.70275","url":null,"abstract":"<div>\u0000 \u0000 <p>Gastrointestinal disorders include diseases that negatively affect people's daily life and carry the risk of cancer. Therefore, accurate and early diagnosis of these diseases is important for treatment process of patients. Deep learning architectures, which have achieved significant success in medical image analysis, are effectively used in early diagnosis and diagnosis systems. Therefore, in this study, a new approach that achieves higher accuracy in the detection of gastrointestinal diseases by combining DenseNet and Vision Transformer models with stacking ensemble is proposed. As a result of the experiments, the proposed model achieved 99.06% accuracy in a single test and 98.64% accuracy on mean as a result of 5-fold cross-validation. The proposed approach shows promising accuracy and reliability as evidenced by the results of experiments on the KvasirV2 dataset, and has the potential to be an effective method for the detection of gastrointestinal diseases. To improve model interpretability, the Explainable AI technique Grad-CAM and attention map visualizations were used, allowing visual justification of the model's predictions and highlighting clinically relevant regions in endoscopic images. The model obtained by combining DenseNet and Vision Transformer models with the stacking ensemble method is expected to be an example for future studies in the field of health and image processing, especially gastrointestinal diseases.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145885036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Attention-Guided Deep Learning Approach for Classifying 39 Skin Lesion Types 基于注意力引导的深度学习方法对39种皮肤病变类型进行分类
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-22 DOI: 10.1002/ima.70269
Sauda Adiv Hanum, Ashim Dey, Muhammad Ashad Kabir

The skin, the largest organ of the human body, is vulnerable to numerous pathological conditions collectively referred to as skin lesions, encompassing a wide spectrum of dermatoses. Diagnosing these lesions remains challenging for medical practitioners due to their subtle visual differences, many of which are imperceptible to the naked eye. While not all lesions are malignant, some serve as early indicators of serious diseases such as skin cancer, emphasizing the urgent need for accurate and timely diagnostic tools. This study advances dermatological diagnostics by curating a comprehensive and balanced dataset containing 9360 dermoscopic and clinical images across 39 lesion categories, synthesized from five publicly available datasets. Five state-of-the-art deep learning architectures—MobileNetV2, Xception, InceptionV3, EfficientNetB1, and Vision Transformer (ViT)—were systematically evaluated on this dataset. To enhance model precision and robustness, Efficient Channel Attention (ECA) and Convolutional Block Attention Module (CBAM) mechanisms were integrated into these architectures. Extensive evaluation across multiple performance metrics demonstrated that the Vision Transformer with CBAM achieved the best results, with 93.46% accuracy, 94% precision, 93% recall, 93% F1-score, and 93.67% specificity. These findings highlight the effectiveness of attention-guided Vision Transformers in addressing complex, large-scale, multi-class skin lesion classification. By combining dataset diversity with advanced attention mechanisms, the proposed framework provides a reliable and interpretable tool to assist medical professionals in accurate and efficient lesion diagnosis, thereby contributing to improved clinical decision-making and patient outcomes.

皮肤是人体最大的器官,容易受到许多病理状况的影响,统称为皮肤病变,包括广泛的皮肤病。诊断这些病变仍然具有挑战性的医疗从业者,由于他们的细微的视觉差异,其中许多是难以察觉的肉眼。虽然并非所有病变都是恶性的,但有些病变可以作为皮肤癌等严重疾病的早期指标,强调迫切需要准确和及时的诊断工具。本研究通过整理一个全面而平衡的数据集,包括39个病变类别的9360张皮肤镜和临床图像,从5个公开的数据集合成,从而推进了皮肤科诊断。五个最先进的深度学习架构——mobilenetv2、Xception、InceptionV3、EfficientNetB1和Vision Transformer (ViT)——在这个数据集上进行了系统评估。为了提高模型的精度和鲁棒性,将高效通道注意(ECA)和卷积块注意模块(CBAM)机制集成到这些体系结构中。对多个性能指标的广泛评估表明,带有CBAM的Vision Transformer达到了最佳效果,准确率为93.46%,精密度为94%,召回率为93%,f1评分为93%,特异性为93.67%。这些发现强调了注意引导视觉变形在处理复杂、大规模、多类别皮肤病变分类中的有效性。通过将数据集多样性与先进的注意机制相结合,所提出的框架提供了一个可靠且可解释的工具,以帮助医疗专业人员准确有效地诊断病变,从而有助于改善临床决策和患者预后。
{"title":"An Attention-Guided Deep Learning Approach for Classifying 39 Skin Lesion Types","authors":"Sauda Adiv Hanum,&nbsp;Ashim Dey,&nbsp;Muhammad Ashad Kabir","doi":"10.1002/ima.70269","DOIUrl":"https://doi.org/10.1002/ima.70269","url":null,"abstract":"<div>\u0000 \u0000 <p>The skin, the largest organ of the human body, is vulnerable to numerous pathological conditions collectively referred to as skin lesions, encompassing a wide spectrum of dermatoses. Diagnosing these lesions remains challenging for medical practitioners due to their subtle visual differences, many of which are imperceptible to the naked eye. While not all lesions are malignant, some serve as early indicators of serious diseases such as skin cancer, emphasizing the urgent need for accurate and timely diagnostic tools. This study advances dermatological diagnostics by curating a comprehensive and balanced dataset containing 9360 dermoscopic and clinical images across 39 lesion categories, synthesized from five publicly available datasets. Five state-of-the-art deep learning architectures—MobileNetV2, Xception, InceptionV3, EfficientNetB1, and Vision Transformer (ViT)—were systematically evaluated on this dataset. To enhance model precision and robustness, Efficient Channel Attention (ECA) and Convolutional Block Attention Module (CBAM) mechanisms were integrated into these architectures. Extensive evaluation across multiple performance metrics demonstrated that the Vision Transformer with CBAM achieved the best results, with 93.46% accuracy, 94% precision, 93% recall, 93% F1-score, and 93.67% specificity. These findings highlight the effectiveness of attention-guided Vision Transformers in addressing complex, large-scale, multi-class skin lesion classification. By combining dataset diversity with advanced attention mechanisms, the proposed framework provides a reliable and interpretable tool to assist medical professionals in accurate and efficient lesion diagnosis, thereby contributing to improved clinical decision-making and patient outcomes.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145825131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comprehensive Multitask Ensemble Segmentation and Clinical Interpretation of Pancreatic and Peripancreatic Anatomy With Radiomics and Deep Learning Features 基于放射组学和深度学习特征的胰腺和胰腺周围解剖的综合多任务集成分割和临床解释
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-20 DOI: 10.1002/ima.70270
Ming Jiang, Jinye Hu, Jie Zheng, Jin Wang, Xiaohui Ye

To develop and validate a multitask deep learning framework for the simultaneous segmentation and clinical classification of pancreatic and peripancreatic anatomical structures in contrast-enhanced CT imaging, enabling robust, automated diagnostic assessment and TNM staging. In this retrospective multicenter study, 3019 contrast-enhanced abdominal CT scans from patients with confirmed or suspected pancreatic disease were analyzed. Six anatomical structures were manually annotated: tumor, parenchyma, pancreatic duct, common bile duct, peripancreatic veins, and arteries. An ensemble model combining nnU-Net, TransUNet, and Swin-UNet was trained for segmentation. Post-segmentation, 215 radiomic and 2560 deep features were extracted and filtered via ICC, correlation, and harmonization procedures. Feature selection was performed using LASSO, MI, and ANOVA. Clinical classification was conducted using XGBoost, MLP, and TabTransformer models. Performance was evaluated through five-fold cross-validation and tested on independent internal and external datasets. The ensemble model achieved high segmentation accuracy (mean DSC: 0.89–0.94 across structures) and superior boundary precision (HD95: < 3 mm). For classification tasks, the best-performing models attained AUCs of 95.5% for tumor malignancy, 94.7% for parenchymal condition, 94.8% for ductal status, and 94.1% for vessel invasion. Feature reproducibility was confirmed with ICC ≥ 0.75 for 198 radiomic and 2112 deep features. External validation confirmed high accuracy and generalizability, with minimal performance degradation across clinical sites. Our multitask AI framework offers comprehensive and clinically actionable insights from CT imaging, combining precise anatomical segmentation with diagnostic classification. The system supports automated TNM staging and demonstrates strong potential for clinical integration. Future studies should explore multimodal imaging, longitudinal data, and genomics integration to further expand its diagnostic and prognostic capabilities.

开发并验证一个多任务深度学习框架,用于对比增强CT成像中胰腺和胰腺周围解剖结构的同时分割和临床分类,实现稳健的自动诊断评估和TNM分期。在这项回顾性多中心研究中,分析了3019例确诊或疑似胰腺疾病患者的腹部CT增强扫描。人工注释6个解剖结构:肿瘤、实质、胰管、胆总管、胰周静脉和动脉。结合nnU-Net、TransUNet和swan - unet的集成模型进行分割训练。分割后,通过ICC、相关和协调程序提取和过滤215个放射性特征和2560个深度特征。特征选择采用LASSO、MI和方差分析。采用XGBoost、MLP、TabTransformer模型进行临床分型。通过五倍交叉验证评估性能,并在独立的内部和外部数据集上进行测试。该集成模型具有较高的分割精度(跨结构平均DSC: 0.89-0.94)和较好的边界精度(HD95: 3 mm)。对于分类任务,表现最好的模型的肿瘤恶性auc为95.5%,实质情况为94.7%,导管状态为94.8%,血管侵犯为94.1%。198个放射学特征和2112个深部特征的ICC≥0.75证实了特征的再现性。外部验证证实了高准确性和通用性,在临床场所的性能下降最小。我们的多任务人工智能框架从CT成像中提供全面和临床可操作的见解,将精确的解剖分割与诊断分类相结合。该系统支持自动TNM分期,并显示出强大的临床整合潜力。未来的研究应探索多模式成像、纵向数据和基因组学整合,以进一步扩大其诊断和预后能力。
{"title":"Comprehensive Multitask Ensemble Segmentation and Clinical Interpretation of Pancreatic and Peripancreatic Anatomy With Radiomics and Deep Learning Features","authors":"Ming Jiang,&nbsp;Jinye Hu,&nbsp;Jie Zheng,&nbsp;Jin Wang,&nbsp;Xiaohui Ye","doi":"10.1002/ima.70270","DOIUrl":"https://doi.org/10.1002/ima.70270","url":null,"abstract":"<p>To develop and validate a multitask deep learning framework for the simultaneous segmentation and clinical classification of pancreatic and peripancreatic anatomical structures in contrast-enhanced CT imaging, enabling robust, automated diagnostic assessment and TNM staging. In this retrospective multicenter study, 3019 contrast-enhanced abdominal CT scans from patients with confirmed or suspected pancreatic disease were analyzed. Six anatomical structures were manually annotated: tumor, parenchyma, pancreatic duct, common bile duct, peripancreatic veins, and arteries. An ensemble model combining nnU-Net, TransUNet, and Swin-UNet was trained for segmentation. Post-segmentation, 215 radiomic and 2560 deep features were extracted and filtered via ICC, correlation, and harmonization procedures. Feature selection was performed using LASSO, MI, and ANOVA. Clinical classification was conducted using XGBoost, MLP, and TabTransformer models. Performance was evaluated through five-fold cross-validation and tested on independent internal and external datasets. The ensemble model achieved high segmentation accuracy (mean DSC: 0.89–0.94 across structures) and superior boundary precision (HD95: &lt; 3 mm). For classification tasks, the best-performing models attained AUCs of 95.5% for tumor malignancy, 94.7% for parenchymal condition, 94.8% for ductal status, and 94.1% for vessel invasion. Feature reproducibility was confirmed with ICC ≥ 0.75 for 198 radiomic and 2112 deep features. External validation confirmed high accuracy and generalizability, with minimal performance degradation across clinical sites. Our multitask AI framework offers comprehensive and clinically actionable insights from CT imaging, combining precise anatomical segmentation with diagnostic classification. The system supports automated TNM staging and demonstrates strong potential for clinical integration. Future studies should explore multimodal imaging, longitudinal data, and genomics integration to further expand its diagnostic and prognostic capabilities.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.70270","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145824901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Deep Learning-Based DenseEchoNet Framework With eXplainable Artificial Intelligence for Accurate and Early Heart Disease Prediction 基于深度学习的密集回声网络框架与可解释的人工智能,用于准确和早期的心脏病预测
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-19 DOI: 10.1002/ima.70268
Meghavathu S. S. Nayak, Hussain Syed

Heart disease (HD) is still a major cause of death worldwide, which emphasizes the importance of early and precise prediction. This paper presents DenseEchoNet, a deep learning model that has been optimized with the Gazelle Optimizer Algorithm (GOA). The hybrid HD-ENN technique is used for balanced learning to solve class imbalance and high dimensionality, while squared exponential kernel-based PCA (SEKPCA) effectively reduces dimensionality. DenseEchoNet outperforms current baseline models with accuracies of 0.9795 as well as 0.9785, respectively, when tested on the HDHI and Cleveland datasets. XAI approaches, such as LIME and SHAP, improve model interpretability by offering distinct insights on feature contributions to HD risk. For early HD prediction, this system provides a straightforward, accurate, and efficient solution.

心脏病(HD)仍然是世界范围内死亡的主要原因,这强调了早期和精确预测的重要性。本文介绍了DenseEchoNet,这是一个用Gazelle优化算法(GOA)优化的深度学习模型。混合HD-ENN技术用于平衡学习,解决了类不平衡和高维问题,而基于平方指数核的PCA (SEKPCA)有效地降低了维数。当在HDHI和Cleveland数据集上测试时,DenseEchoNet的精度分别为0.9795和0.9785,优于当前的基线模型。XAI方法,如LIME和SHAP,通过提供对HD风险的特征贡献的独特见解,提高了模型的可解释性。对于HD的早期预测,该系统提供了一种简单、准确、高效的解决方案。
{"title":"A Deep Learning-Based DenseEchoNet Framework With eXplainable Artificial Intelligence for Accurate and Early Heart Disease Prediction","authors":"Meghavathu S. S. Nayak,&nbsp;Hussain Syed","doi":"10.1002/ima.70268","DOIUrl":"https://doi.org/10.1002/ima.70268","url":null,"abstract":"<div>\u0000 \u0000 <p>Heart disease (HD) is still a major cause of death worldwide, which emphasizes the importance of early and precise prediction. This paper presents DenseEchoNet, a deep learning model that has been optimized with the Gazelle Optimizer Algorithm (GOA). The hybrid HD-ENN technique is used for balanced learning to solve class imbalance and high dimensionality, while squared exponential kernel-based PCA (SEKPCA) effectively reduces dimensionality. DenseEchoNet outperforms current baseline models with accuracies of 0.9795 as well as 0.9785, respectively, when tested on the HDHI and Cleveland datasets. XAI approaches, such as LIME and SHAP, improve model interpretability by offering distinct insights on feature contributions to HD risk. For early HD prediction, this system provides a straightforward, accurate, and efficient solution.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145824666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Application of MLEM-TV Algorithm in Diffuse Correlation Tomography Blood Flow Imaging MLEM-TV算法在弥散相关断层血流成像中的应用
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-10 DOI: 10.1002/ima.70267
Zicheng Li, Dalin Cheng, Juanjuan Shen, Xiaojuan Zhang

Diffuse correlation tomography (DCT) reconstructs the motion velocity of scatterers (blood flow index, BFI) within biological tissues by using information from the escaped photons. Given the randomness of a single scatterer motion and the determinism of particle swarm, the motion of scatterers was analyzed for the first time using a probabilistic-statistical method. This study applied the maximum likelihood expectation maximization (MLEM) algorithm for DCT, integrating a total variation (TV) regularization model as a constraint to enhance BFI reconstruction. In simulation, the mean absolute error (MAE) of the cross-shaped anomaly, reconstructed from the noise-free and noisy autocorrelation function g1(τ), was 0.0962 and 0.1831, with the corresponding contrast of 8.25 and 6.42, respectively. For the two-dot anomaly, the MAE was 0.0293 and 0.0452, with the corresponding contrast of 4.07 and 3.08, respectively. In phantom experiments, the contrast of the cross-shaped anomaly was 0.59. For the controllable velocity tubular anomaly, the contrast (1.42, 1.95, and 2.49) is gradually enhanced as the pump speed is elevated. Clinical tests of calf skeletal muscle revealed approximately tenfold higher BFI in the relaxed state than in the cuff occlusion state. The result demonstrates that the MLEM-TV algorithm can be an alternative algorithm for BFI reconstruction, with potential applications for detecting abnormal blood flow perfusion in cerebral, breast, and skeletal muscle pathology.

漫射相关断层扫描(DCT)利用逸出光子的信息重建生物组织内散射体(血流指数,BFI)的运动速度。考虑到单个散射体运动的随机性和粒子群的确定性,首次采用概率统计方法对散射体运动进行了分析。本研究将最大似然期望最大化(MLEM)算法应用于DCT,整合一个总变分(TV)正则化模型作为约束来增强BFI重建。在模拟中,由无噪声和有噪声的自相关函数g1(τ)重建的十字形异常的平均绝对误差(MAE)分别为0.0962和0.1831,对应的对比度分别为8.25和6.42。两点异常的MAE分别为0.0293和0.0452,对应的对比分别为4.07和3.08。在幻像实验中,十字形异常的对比度为0.59。对于可控速度管状异常,随着泵转速的提高,对比值(1.42、1.95、2.49)逐渐增强。小腿骨骼肌的临床试验显示,放松状态下的BFI比袖带闭塞状态高约10倍。结果表明,MLEM-TV算法可以作为BFI重建的替代算法,在检测脑、乳腺和骨骼肌病理异常血流灌注方面具有潜在的应用前景。
{"title":"Application of MLEM-TV Algorithm in Diffuse Correlation Tomography Blood Flow Imaging","authors":"Zicheng Li,&nbsp;Dalin Cheng,&nbsp;Juanjuan Shen,&nbsp;Xiaojuan Zhang","doi":"10.1002/ima.70267","DOIUrl":"https://doi.org/10.1002/ima.70267","url":null,"abstract":"<p>Diffuse correlation tomography (DCT) reconstructs the motion velocity of scatterers (blood flow index, BFI) within biological tissues by using information from the escaped photons. Given the randomness of a single scatterer motion and the determinism of particle swarm, the motion of scatterers was analyzed for the first time using a probabilistic-statistical method. This study applied the maximum likelihood expectation maximization (MLEM) algorithm for DCT, integrating a total variation (TV) regularization model as a constraint to enhance BFI reconstruction. In simulation, the mean absolute error (MAE) of the <i>cross-shaped</i> anomaly, reconstructed from the noise-free and noisy autocorrelation function <i>g</i><sub>1</sub>(<i>τ</i>), was 0.0962 and 0.1831, with the corresponding contrast of 8.25 and 6.42, respectively. For the <i>two-dot</i> anomaly, the MAE was 0.0293 and 0.0452, with the corresponding contrast of 4.07 and 3.08, respectively. In phantom experiments, the contrast of the <i>cross-shaped</i> anomaly was 0.59. For the controllable velocity <i>tubular</i> anomaly, the contrast (1.42, 1.95, and 2.49) is gradually enhanced as the pump speed is elevated. Clinical tests of calf skeletal muscle revealed approximately tenfold higher BFI in the relaxed state than in the cuff occlusion state. The result demonstrates that the MLEM-TV algorithm can be an alternative algorithm for BFI reconstruction, with potential applications for detecting abnormal blood flow perfusion in cerebral, breast, and skeletal muscle pathology.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.70267","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HAI-Net: Skin Lesion Segmentation Using a High-Performance Adaptive Attention and Information Interaction Network HAI-Net:基于高性能自适应注意与信息交互网络的皮肤病变分割
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-10 DOI: 10.1002/ima.70266
Chao Fan, Li Chen, Mengyang Yun, Huijun Zhao, Bincheng Peng

Skin lesion segmentation from dermoscopic images must be done accurately and consistently in order to diagnose diseases and arrange treatments. However, when dealing with issues like fuzzy lesion region boundaries, multiscale features, and notable variations in the lesion region's size, shape, and color, existing methods typically have high computational complexity and large parameter counts. They also frequently suffer from decreased segmentation accuracy due to inadequate capture of local features and global information. In this paper, a lightweight deep learning network based on high-performance adaptive attention is proposed to overcome these issues. Specifically, a deep convolutional neural network is introduced to capture local information. Meanwhile, we create a high-performance adaptive attention feature fusion module (EAAF) that uses dynamic feature selection to achieve adaptive fusion of global information with multiscale local features. Furthermore, we created a reverse dynamic feature fusion module (RDFM) at the decoding stage to efficiently fuse features at various levels while taking into account the integrity and specifics of the lesion region to increase the precision of complex lesion region segmentation. We carried out in-depth tests on three publicly accessible datasets International Skin Imaging Collaboration (ISIC)-2016, ISIC-2018, and PH2 to assess the method's efficacy and contrasted the outcomes with those of the most advanced techniques; the results confirmed that the suggested approach was superior.

为了诊断疾病和安排治疗,必须准确、一致地对皮肤镜图像中的皮肤病变进行分割。然而,在处理模糊的病灶区域边界、多尺度特征以及病灶区域的大小、形状和颜色变化明显等问题时,现有方法通常具有较高的计算复杂度和较大的参数计数。由于对局部特征和全局信息的捕捉不足,它们也经常遭受分割精度下降的困扰。为了克服这些问题,本文提出了一种基于高性能自适应注意力的轻量级深度学习网络。具体来说,引入了深度卷积神经网络来捕获局部信息。同时,构建了高性能的自适应关注特征融合模块(EAAF),利用动态特征选择实现全局信息与多尺度局部特征的自适应融合。此外,我们在解码阶段创建了反向动态特征融合模块(RDFM),在考虑病变区域的完整性和特殊性的同时,有效地融合了各级特征,提高了复杂病变区域分割的精度。我们对国际皮肤成像合作组织(ISIC)-2016、ISIC-2018和PH2三个可公开访问的数据集进行了深入测试,以评估该方法的疗效,并将结果与最先进的技术进行了对比;结果证实了该方法的优越性。
{"title":"HAI-Net: Skin Lesion Segmentation Using a High-Performance Adaptive Attention and Information Interaction Network","authors":"Chao Fan,&nbsp;Li Chen,&nbsp;Mengyang Yun,&nbsp;Huijun Zhao,&nbsp;Bincheng Peng","doi":"10.1002/ima.70266","DOIUrl":"https://doi.org/10.1002/ima.70266","url":null,"abstract":"<div>\u0000 \u0000 <p>Skin lesion segmentation from dermoscopic images must be done accurately and consistently in order to diagnose diseases and arrange treatments. However, when dealing with issues like fuzzy lesion region boundaries, multiscale features, and notable variations in the lesion region's size, shape, and color, existing methods typically have high computational complexity and large parameter counts. They also frequently suffer from decreased segmentation accuracy due to inadequate capture of local features and global information. In this paper, a lightweight deep learning network based on high-performance adaptive attention is proposed to overcome these issues. Specifically, a deep convolutional neural network is introduced to capture local information. Meanwhile, we create a high-performance adaptive attention feature fusion module (EAAF) that uses dynamic feature selection to achieve adaptive fusion of global information with multiscale local features. Furthermore, we created a reverse dynamic feature fusion module (RDFM) at the decoding stage to efficiently fuse features at various levels while taking into account the integrity and specifics of the lesion region to increase the precision of complex lesion region segmentation. We carried out in-depth tests on three publicly accessible datasets International Skin Imaging Collaboration (ISIC)-2016, ISIC-2018, and PH<sup>2</sup> to assess the method's efficacy and contrasted the outcomes with those of the most advanced techniques; the results confirmed that the suggested approach was superior.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750881","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SEA-Net: Dual Attention U-Net for Bleeding Segmentation in Capsule Endoscopy Images SEA-Net:用于胶囊内窥镜图像出血分割的双注意力U-Net
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-10 DOI: 10.1002/ima.70261
Tareque Bashar Ovi, Nomaiya Bashree, Hussain Nyeem, Md Abdul Wahed, Faiaz Hasanuzzaman Rhythm, Disha Chowdhury

Gastrointestinal (GI) bleeding, arising from various conditions, can be critical if untreated. Wireless capsule endoscopy (WCE) is a highly effective method for detecting GI bleeding, offering full visualization of the GI tract. However, the large number of images generated per patient poses challenges for clinicians, leading to prolonged analysis times and increased risk of human error. This emphasizes the need for computer-aided diagnosis systems. In this study, we introduce SEA-Net (Structured Efficient Attention Network), a novel deep learning network for detecting bleeding regions in WCE images. SEA-Net integrates a Convolutional Block Attention Module (CBAM) with long skip connections to enhance gradient flow and improve blood region localization. The EfficientNet-B4 encoder improves feature extraction efficiency and generalizability. A five-fold cross validation demonstrates consistent performance, while generalization tests, including precision-recall curves, ROC curves, and F1 measure, further validate the model's robustness. Minimal performance degradation was observed when the training data was reduced from 80% to 20%. Experimental results show that SEA-Net achieves a Dice score of 93.64% and an IoU score of 88.61% on a publicly available WCE dataset, outperforming state-of-the-art models and highlighting its strong potential for clinical application.

胃肠道(GI)出血,由各种情况引起,如果不治疗,可能会很严重。无线胶囊内窥镜(WCE)是一种非常有效的检测胃肠道出血的方法,提供了胃肠道的全面可视化。然而,每位患者生成的大量图像给临床医生带来了挑战,导致分析时间延长,人为错误的风险增加。这强调了对计算机辅助诊断系统的需求。在本研究中,我们引入了一种新的深度学习网络SEA-Net (Structured Efficient Attention Network),用于检测WCE图像中的出血区域。SEA-Net集成了一个带有长跳跃连接的卷积块注意模块(CBAM),以增强梯度流动和改善血液区域定位。高效网- b4编码器提高了特征提取效率和通用性。五重交叉验证显示了一致的性能,而包括精度-召回率曲线、ROC曲线和F1测量在内的泛化测试进一步验证了模型的稳健性。当训练数据从80%减少到20%时,观察到的性能下降最小。实验结果表明,SEA-Net在公开可用的WCE数据集上的Dice得分为93.64%,IoU得分为88.61%,优于目前最先进的模型,显示了其强大的临床应用潜力。
{"title":"SEA-Net: Dual Attention U-Net for Bleeding Segmentation in Capsule Endoscopy Images","authors":"Tareque Bashar Ovi,&nbsp;Nomaiya Bashree,&nbsp;Hussain Nyeem,&nbsp;Md Abdul Wahed,&nbsp;Faiaz Hasanuzzaman Rhythm,&nbsp;Disha Chowdhury","doi":"10.1002/ima.70261","DOIUrl":"https://doi.org/10.1002/ima.70261","url":null,"abstract":"<div>\u0000 \u0000 <p>Gastrointestinal (GI) bleeding, arising from various conditions, can be critical if untreated. Wireless capsule endoscopy (WCE) is a highly effective method for detecting GI bleeding, offering full visualization of the GI tract. However, the large number of images generated per patient poses challenges for clinicians, leading to prolonged analysis times and increased risk of human error. This emphasizes the need for computer-aided diagnosis systems. In this study, we introduce SEA-Net (<b>S</b>tructured <b>E</b>fficient <b>A</b>ttention <b>Net</b>work), a novel deep learning network for detecting bleeding regions in WCE images. SEA-Net integrates a Convolutional Block Attention Module (CBAM) with long skip connections to enhance gradient flow and improve blood region localization. The EfficientNet-B4 encoder improves feature extraction efficiency and generalizability. A five-fold cross validation demonstrates consistent performance, while generalization tests, including precision-recall curves, ROC curves, and F1 measure, further validate the model's robustness. Minimal performance degradation was observed when the training data was reduced from 80% to 20%. Experimental results show that SEA-Net achieves a Dice score of 93.64% and an IoU score of 88.61% on a publicly available WCE dataset, outperforming state-of-the-art models and highlighting its strong potential for clinical application.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Deep Fuzzy Inference System for Interpretable Multi-Class Heart Disease Risk Prediction 可解释多类别心脏病风险预测的深度模糊推理系统
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-10 DOI: 10.1002/ima.70264
S. Ramasami, P. Uma Maheswari

Heart disease remains a leading global health concern, necessitating accurate and interpretable risk prediction models for effective clinical decision-making. Accurate heart disease risk prediction is crucial for preventive healthcare, yet traditional machine learning models often struggle with the inherent uncertainty and nonlinear patterns in medical data. While deep neural networks (DNNs) excel at feature extraction, their lack of interpretability limits clinical utility, whereas fuzzy inference systems (FIS) offer transparency but lack hierarchical learning capabilities. To bridge this gap, we propose a novel Deep Fuzzy Inference System (DFIS) that integrates DNNs and FIS into a unified architecture, combining the strengths of both approaches. The DFIS leverages a weighted fusion mechanism to combine probabilistic outputs from an Adam Cuckoo Search-optimized DNN and a trapezoidal membership-based FIS, enabling simultaneous high accuracy and interpretability. The performance is evaluated on the Cleveland heart disease dataset; the DFIS achieves 97.2% accuracy, outperforming a standalone DNN (95.4%) and ANFIS (91.7%) under the same experimental conditions while providing clinically actionable risk stratification into normal, less critical, and very critical categories.

心脏病仍然是全球主要的健康问题,需要准确和可解释的风险预测模型来进行有效的临床决策。准确的心脏病风险预测对于预防性医疗保健至关重要,但传统的机器学习模型经常与医疗数据中固有的不确定性和非线性模式作斗争。虽然深度神经网络(dnn)擅长特征提取,但其缺乏可解释性限制了临床应用,而模糊推理系统(FIS)提供透明度,但缺乏分层学习能力。为了弥补这一差距,我们提出了一种新的深度模糊推理系统(DFIS),它将dnn和FIS集成到一个统一的架构中,结合了两种方法的优势。DFIS利用加权融合机制,将亚当布谷鸟搜索优化的深度神经网络和基于梯形隶属度的FIS的概率输出结合起来,同时实现高精度和可解释性。在克利夫兰心脏病数据集上对性能进行评估;在相同的实验条件下,DFIS达到97.2%的准确率,优于独立DNN(95.4%)和ANFIS(91.7%),同时提供临床可操作的风险分层,分为正常、不太严重和非常严重的类别。
{"title":"A Deep Fuzzy Inference System for Interpretable Multi-Class Heart Disease Risk Prediction","authors":"S. Ramasami,&nbsp;P. Uma Maheswari","doi":"10.1002/ima.70264","DOIUrl":"https://doi.org/10.1002/ima.70264","url":null,"abstract":"<div>\u0000 \u0000 <p>Heart disease remains a leading global health concern, necessitating accurate and interpretable risk prediction models for effective clinical decision-making. Accurate heart disease risk prediction is crucial for preventive healthcare, yet traditional machine learning models often struggle with the inherent uncertainty and nonlinear patterns in medical data. While deep neural networks (DNNs) excel at feature extraction, their lack of interpretability limits clinical utility, whereas fuzzy inference systems (FIS) offer transparency but lack hierarchical learning capabilities. To bridge this gap, we propose a novel Deep Fuzzy Inference System (DFIS) that integrates DNNs and FIS into a unified architecture, combining the strengths of both approaches. The DFIS leverages a weighted fusion mechanism to combine probabilistic outputs from an Adam Cuckoo Search-optimized DNN and a trapezoidal membership-based FIS, enabling simultaneous high accuracy and interpretability. The performance is evaluated on the Cleveland heart disease dataset; the DFIS achieves 97.2% accuracy, outperforming a standalone DNN (95.4%) and ANFIS (91.7%) under the same experimental conditions while providing clinically actionable risk stratification into normal, less critical, and very critical categories.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Lung Disease Diagnosis: A High Performance Hybrid Deep Learning Framework for Multi-Class Chest X-Ray Analysis 增强肺部疾病诊断:用于多类别胸部x射线分析的高性能混合深度学习框架
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-09 DOI: 10.1002/ima.70262
Tolga Saim Bascetin, Ibrahim Emiroglu

This study presents a high performance hybrid deep learning model for the classification of 14 lung diseases using chest X-ray (CXR) images. Manual evaluation of CXR images is labor-intensive and prone to human error. Therefore, automated systems are required to improve diagnostic accuracy and efficiency. Our model integrates ResNet18 and EfficientNet-V2-S architectures, combining residual connections with efficient scaling to achieve high accuracy while maintaining computational efficiency. Trained on the NIH ChestX-ray14 dataset, comprising 112 120 images across 14 disease classes, the model mitigates class imbalances with extensive data augmentation techniques. Achieving an impressive average AUC of 0.872, the model outperforms previous approaches. This performance was enhanced by a refined, anatomically-aware data augmentation strategy that improved the model's robustness and clinical relevance, particularly in challenging disease categories such as Pneumothorax, Emphysema, and Hernia. To further validate its generalizability, the proposed model was tested on three additional datasets for pneumonia, COVID-19, and tuberculosis. The results demonstrate superior performance, achieving an accuracy of 0.958, F1 score of 0.944, and ROC AUC of 0.989 for pneumonia; an accuracy of 0.974, F1 score of 0.969, and ROC AUC of 0.995 for COVID-19; and an accuracy of 0.999, F1 score of 0.999, and ROC AUC of 0.999 for tuberculosis. These outstanding results confirm the robustness and clinical applicability of the model across diverse datasets. This research introduces a reliable and efficient diagnostic tool that enhances the potential of automated lung disease classification. By alleviating radiologists' workload and promoting timely, accurate diagnostic outcomes, the model contributes significantly to medical imaging applications and demonstrates its capacity for practical use in real-world clinical settings.

本研究提出了一种高性能混合深度学习模型,用于使用胸部x射线(CXR)图像对14种肺部疾病进行分类。人工评估CXR图像是一项劳动密集型工作,而且容易出现人为错误。因此,需要自动化系统来提高诊断的准确性和效率。我们的模型集成了ResNet18和EfficientNet-V2-S架构,将剩余连接与高效缩放相结合,在保持计算效率的同时实现高精度。该模型在NIH ChestX-ray14数据集上进行训练,该数据集包含14种疾病类别的112 - 120张图像,通过广泛的数据增强技术减轻了类别不平衡。该模型获得了令人印象深刻的平均AUC 0.872,优于以前的方法。这一性能通过一种精细的、具有解剖学意识的数据增强策略得到了增强,该策略提高了模型的稳健性和临床相关性,特别是在气胸、肺气肿和疝气等具有挑战性的疾病类别中。为了进一步验证其普遍性,在肺炎、COVID-19和结核病的另外三个数据集上测试了所提出的模型。结果显示,该方法对肺炎的诊断准确率为0.958,F1评分为0.944,ROC AUC为0.989;COVID-19的准确率为0.974,F1评分为0.969,ROC AUC为0.995;诊断肺结核的准确率为0.999,F1评分为0.999,ROC AUC为0.999。这些突出的结果证实了该模型在不同数据集上的稳健性和临床适用性。本研究介绍了一种可靠、高效的诊断工具,提高了肺部疾病自动分类的潜力。通过减轻放射科医生的工作量,促进及时、准确的诊断结果,该模型为医学成像应用做出了重大贡献,并展示了其在现实世界临床环境中的实际应用能力。
{"title":"Enhancing Lung Disease Diagnosis: A High Performance Hybrid Deep Learning Framework for Multi-Class Chest X-Ray Analysis","authors":"Tolga Saim Bascetin,&nbsp;Ibrahim Emiroglu","doi":"10.1002/ima.70262","DOIUrl":"https://doi.org/10.1002/ima.70262","url":null,"abstract":"<div>\u0000 \u0000 <p>This study presents a high performance hybrid deep learning model for the classification of 14 lung diseases using chest X-ray (CXR) images. Manual evaluation of CXR images is labor-intensive and prone to human error. Therefore, automated systems are required to improve diagnostic accuracy and efficiency. Our model integrates ResNet18 and EfficientNet-V2-S architectures, combining residual connections with efficient scaling to achieve high accuracy while maintaining computational efficiency. Trained on the NIH ChestX-ray14 dataset, comprising 112 120 images across 14 disease classes, the model mitigates class imbalances with extensive data augmentation techniques. Achieving an impressive average AUC of 0.872, the model outperforms previous approaches. This performance was enhanced by a refined, anatomically-aware data augmentation strategy that improved the model's robustness and clinical relevance, particularly in challenging disease categories such as Pneumothorax, Emphysema, and Hernia. To further validate its generalizability, the proposed model was tested on three additional datasets for pneumonia, COVID-19, and tuberculosis. The results demonstrate superior performance, achieving an accuracy of 0.958, <i>F</i>1 score of 0.944, and ROC AUC of 0.989 for pneumonia; an accuracy of 0.974, <i>F</i>1 score of 0.969, and ROC AUC of 0.995 for COVID-19; and an accuracy of 0.999, <i>F</i>1 score of 0.999, and ROC AUC of 0.999 for tuberculosis. These outstanding results confirm the robustness and clinical applicability of the model across diverse datasets. This research introduces a reliable and efficient diagnostic tool that enhances the potential of automated lung disease classification. By alleviating radiologists' workload and promoting timely, accurate diagnostic outcomes, the model contributes significantly to medical imaging applications and demonstrates its capacity for practical use in real-world clinical settings.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1