Automated classification of chest X-rays: a deep learning approach with attention mechanisms.

IF 3.2 3区 医学 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING BMC Medical Imaging Pub Date : 2025-03-04 DOI:10.1186/s12880-025-01604-5
Burcu Oltu, Selda Güney, Seniha Esen Yuksel, Berna Dengiz
{"title":"Automated classification of chest X-rays: a deep learning approach with attention mechanisms.","authors":"Burcu Oltu, Selda Güney, Seniha Esen Yuksel, Berna Dengiz","doi":"10.1186/s12880-025-01604-5","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Pulmonary diseases such as COVID-19 and pneumonia, are life-threatening conditions, that require prompt and accurate diagnosis for effective treatment. Chest X-ray (CXR) has become the most common alternative method for detecting pulmonary diseases such as COVID-19, pneumonia, and lung opacity due to their availability, cost-effectiveness, and ability to facilitate comparative analysis. However, the interpretation of CXRs is a challenging task.</p><p><strong>Methods: </strong>This study presents an automated deep learning (DL) model that outperforms multiple state-of-the-art methods in diagnosing COVID-19, Lung Opacity, and Viral Pneumonia. Using a dataset of 21,165 CXRs, the proposed framework introduces a seamless combination of the Vision Transformer (ViT) for capturing long-range dependencies, DenseNet201 for powerful feature extraction, and global average pooling (GAP) for retaining critical spatial details. This combination results in a robust classification system, achieving remarkable accuracy.</p><p><strong>Results: </strong>The proposed methodology delivers outstanding results across all categories: achieving 99.4% accuracy and an F1-score of 98.43% for COVID-19, 96.45% accuracy and an F1-score of 93.64% for Lung Opacity, 99.63% accuracy and an F1-score of 97.05% for Viral Pneumonia, and 95.97% accuracy with an F1-score of 95.87% for Normal subjects.</p><p><strong>Conclusion: </strong>The proposed framework achieves a remarkable overall accuracy of 97.87%, surpassing several state-of-the-art methods with reproducible and objective outcomes. To ensure robustness and minimize variability in train-test splits, our study employs five-fold cross-validation, providing reliable and consistent performance evaluation. For transparency and to facilitate future comparisons, the specific training and testing splits have been made publicly accessible. Furthermore, Grad-CAM-based visualizations are integrated to enhance the interpretability of the model, offering valuable insights into its decision-making process. This innovative framework not only boosts classification accuracy but also sets a new benchmark in CXR-based disease diagnosis.</p>","PeriodicalId":9020,"journal":{"name":"BMC Medical Imaging","volume":"25 1","pages":"71"},"PeriodicalIF":3.2000,"publicationDate":"2025-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11877751/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"BMC Medical Imaging","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1186/s12880-025-01604-5","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Pulmonary diseases such as COVID-19 and pneumonia, are life-threatening conditions, that require prompt and accurate diagnosis for effective treatment. Chest X-ray (CXR) has become the most common alternative method for detecting pulmonary diseases such as COVID-19, pneumonia, and lung opacity due to their availability, cost-effectiveness, and ability to facilitate comparative analysis. However, the interpretation of CXRs is a challenging task.

Methods: This study presents an automated deep learning (DL) model that outperforms multiple state-of-the-art methods in diagnosing COVID-19, Lung Opacity, and Viral Pneumonia. Using a dataset of 21,165 CXRs, the proposed framework introduces a seamless combination of the Vision Transformer (ViT) for capturing long-range dependencies, DenseNet201 for powerful feature extraction, and global average pooling (GAP) for retaining critical spatial details. This combination results in a robust classification system, achieving remarkable accuracy.

Results: The proposed methodology delivers outstanding results across all categories: achieving 99.4% accuracy and an F1-score of 98.43% for COVID-19, 96.45% accuracy and an F1-score of 93.64% for Lung Opacity, 99.63% accuracy and an F1-score of 97.05% for Viral Pneumonia, and 95.97% accuracy with an F1-score of 95.87% for Normal subjects.

Conclusion: The proposed framework achieves a remarkable overall accuracy of 97.87%, surpassing several state-of-the-art methods with reproducible and objective outcomes. To ensure robustness and minimize variability in train-test splits, our study employs five-fold cross-validation, providing reliable and consistent performance evaluation. For transparency and to facilitate future comparisons, the specific training and testing splits have been made publicly accessible. Furthermore, Grad-CAM-based visualizations are integrated to enhance the interpretability of the model, offering valuable insights into its decision-making process. This innovative framework not only boosts classification accuracy but also sets a new benchmark in CXR-based disease diagnosis.

Abstract Image

Abstract Image

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
胸部x光片的自动分类:一种具有注意机制的深度学习方法。
背景:COVID-19和肺炎等肺部疾病是危及生命的疾病,需要及时准确诊断以进行有效治疗。胸部x光(CXR)由于其可获得性、成本效益和促进比较分析的能力,已成为检测COVID-19、肺炎和肺混浊等肺部疾病最常用的替代方法。然而,对cxr的解释是一项具有挑战性的任务。方法:本研究提出了一种自动深度学习(DL)模型,该模型在诊断COVID-19、肺不透明和病毒性肺炎方面优于多种最先进的方法。使用21,165个cxr数据集,提出的框架引入了视觉转换器(ViT)的无缝组合,用于捕获远程依赖关系,DenseNet201用于强大的特征提取,以及全球平均池(GAP)用于保留关键空间细节。这种组合产生了一个健壮的分类系统,实现了显著的准确性。结果:所提出的方法在所有类别中都取得了出色的结果:COVID-19的准确率为99.4%,f1评分为98.43%,肺不透明的准确率为96.45%,f1评分为93.64%,病毒性肺炎的准确率为99.63%,f1评分为97.05%,正常受试者的准确率为95.97%,f1评分为95.87%。结论:该框架的总体准确率为97.87%,优于几种最先进的方法,具有可重复性和客观的结果。为了确保稳健性和最小化训练测试分割的可变性,我们的研究采用五倍交叉验证,提供可靠和一致的性能评估。为了提高透明度和便于将来的比较,具体的训练和测试划分已经公开。此外,还集成了基于grad - cam的可视化,以增强模型的可解释性,为其决策过程提供有价值的见解。这一创新框架不仅提高了分类准确率,而且为基于cxr的疾病诊断树立了新的标杆。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
BMC Medical Imaging
BMC Medical Imaging RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING-
CiteScore
4.60
自引率
3.70%
发文量
198
审稿时长
27 weeks
期刊介绍: BMC Medical Imaging is an open access journal publishing original peer-reviewed research articles in the development, evaluation, and use of imaging techniques and image processing tools to diagnose and manage disease.
期刊最新文献
Interpretable machine learning models based on CT radiomics for predicting chemoradiotherapy response in rectal cancer. Sonographic features associated with ischemic stroke in patients with extracranial internal carotid artery dissection: a single-center exploratory study. A dynamic nomogram for predicting primary intraoperative brain bulge in patients with traumatic acute subdural hematoma. Comparing YOLO and U-net deep learning algorithms in chronic wound image segmentation. Nomogram combining ultrasound radiomics and clinical factors for predicting placenta accreta spectrum in patients with placenta previa: a two-center study.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1