X-Brain:使用鲁棒深度注意 CNN 对脑肿瘤进行可解释识别

IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Biomedical Signal Processing and Control Pub Date : 2024-10-16 DOI:10.1016/j.bspc.2024.106988
Moshiur Rahman Tonmoy , Md. Atik Shams , Md. Akhtaruzzaman Adnan , M.F. Mridha , Mejdl Safran , Sultan Alfarhood , Dunren Che
{"title":"X-Brain:使用鲁棒深度注意 CNN 对脑肿瘤进行可解释识别","authors":"Moshiur Rahman Tonmoy ,&nbsp;Md. Atik Shams ,&nbsp;Md. Akhtaruzzaman Adnan ,&nbsp;M.F. Mridha ,&nbsp;Mejdl Safran ,&nbsp;Sultan Alfarhood ,&nbsp;Dunren Che","doi":"10.1016/j.bspc.2024.106988","DOIUrl":null,"url":null,"abstract":"<div><div>Automated brain tumor recognition is crucial for swift diagnosis and treatment in healthcare, enhancing patient survival rates but manual recognition of tumor types is time-consuming and resource-intensive. Over the past few years, researchers have proposed various Deep Learning (DL) methods to automate the recognition process over the past years. However, these approaches often lack convincing accuracy and rely on datasets consisting of limited samples, raising concerns regarding real-world efficacy and reliability. Furthermore, the decisions made by black-box AI models in healthcare, where lives are at stake, require proper decision explainability. To address these issues, we propose a robust and explainable deep CNN-based method for effective recognition of brain tumor types, attaining state-of-the-art accuracies of 99.81%, 99.55%, and 99.30% on the training, validation, and test sets, respectively, surpassing both the previous works and baseline models. Moreover, we employed three Explainable AI techniques: Grad-CAM, Grad-CAM++, and Score-CAM for explainability analysis, contributing towards the development of trustworthy and reliable automation of healthcare diagnosis.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":null,"pages":null},"PeriodicalIF":4.9000,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"X-Brain: Explainable recognition of brain tumors using robust deep attention CNN\",\"authors\":\"Moshiur Rahman Tonmoy ,&nbsp;Md. Atik Shams ,&nbsp;Md. Akhtaruzzaman Adnan ,&nbsp;M.F. Mridha ,&nbsp;Mejdl Safran ,&nbsp;Sultan Alfarhood ,&nbsp;Dunren Che\",\"doi\":\"10.1016/j.bspc.2024.106988\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Automated brain tumor recognition is crucial for swift diagnosis and treatment in healthcare, enhancing patient survival rates but manual recognition of tumor types is time-consuming and resource-intensive. Over the past few years, researchers have proposed various Deep Learning (DL) methods to automate the recognition process over the past years. However, these approaches often lack convincing accuracy and rely on datasets consisting of limited samples, raising concerns regarding real-world efficacy and reliability. Furthermore, the decisions made by black-box AI models in healthcare, where lives are at stake, require proper decision explainability. To address these issues, we propose a robust and explainable deep CNN-based method for effective recognition of brain tumor types, attaining state-of-the-art accuracies of 99.81%, 99.55%, and 99.30% on the training, validation, and test sets, respectively, surpassing both the previous works and baseline models. Moreover, we employed three Explainable AI techniques: Grad-CAM, Grad-CAM++, and Score-CAM for explainability analysis, contributing towards the development of trustworthy and reliable automation of healthcare diagnosis.</div></div>\",\"PeriodicalId\":55362,\"journal\":{\"name\":\"Biomedical Signal Processing and Control\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":4.9000,\"publicationDate\":\"2024-10-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Biomedical Signal Processing and Control\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1746809424010462\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, BIOMEDICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomedical Signal Processing and Control","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1746809424010462","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0

摘要

脑肿瘤的自动识别对医疗保健领域的快速诊断和治疗至关重要,可提高患者的存活率,但人工识别肿瘤类型耗时且耗费资源。在过去几年中,研究人员提出了各种深度学习(DL)方法来实现识别过程的自动化。然而,这些方法往往缺乏令人信服的准确性,并且依赖于由有限样本组成的数据集,从而引发了人们对真实世界的有效性和可靠性的担忧。此外,黑盒子人工智能模型在医疗保健领域做出的决策关系到生命安全,因此需要适当的决策可解释性。为了解决这些问题,我们提出了一种基于可解释深度 CNN 的稳健方法,用于有效识别脑肿瘤类型,在训练集、验证集和测试集上的准确率分别达到 99.81%、99.55% 和 99.30%,超越了之前的研究成果和基线模型。此外,我们还采用了三种可解释人工智能技术:此外,我们还采用了三种可解释人工智能技术:Grad-CAM、Grad-CAM++ 和 Score-CAM 进行可解释性分析,为开发可信、可靠的医疗诊断自动化做出了贡献。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
X-Brain: Explainable recognition of brain tumors using robust deep attention CNN
Automated brain tumor recognition is crucial for swift diagnosis and treatment in healthcare, enhancing patient survival rates but manual recognition of tumor types is time-consuming and resource-intensive. Over the past few years, researchers have proposed various Deep Learning (DL) methods to automate the recognition process over the past years. However, these approaches often lack convincing accuracy and rely on datasets consisting of limited samples, raising concerns regarding real-world efficacy and reliability. Furthermore, the decisions made by black-box AI models in healthcare, where lives are at stake, require proper decision explainability. To address these issues, we propose a robust and explainable deep CNN-based method for effective recognition of brain tumor types, attaining state-of-the-art accuracies of 99.81%, 99.55%, and 99.30% on the training, validation, and test sets, respectively, surpassing both the previous works and baseline models. Moreover, we employed three Explainable AI techniques: Grad-CAM, Grad-CAM++, and Score-CAM for explainability analysis, contributing towards the development of trustworthy and reliable automation of healthcare diagnosis.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Biomedical Signal Processing and Control
Biomedical Signal Processing and Control 工程技术-工程:生物医学
CiteScore
9.80
自引率
13.70%
发文量
822
审稿时长
4 months
期刊介绍: Biomedical Signal Processing and Control aims to provide a cross-disciplinary international forum for the interchange of information on research in the measurement and analysis of signals and images in clinical medicine and the biological sciences. Emphasis is placed on contributions dealing with the practical, applications-led research on the use of methods and devices in clinical diagnosis, patient monitoring and management. Biomedical Signal Processing and Control reflects the main areas in which these methods are being used and developed at the interface of both engineering and clinical science. The scope of the journal is defined to include relevant review papers, technical notes, short communications and letters. Tutorial papers and special issues will also be published.
期刊最新文献
Automated pericardium segmentation and epicardial adipose tissue quantification from computed tomography images A design of computational stochastic framework for the mathematical severe acute respiratory syndrome coronavirus model Topological feature search method for multichannel EEG: Application in ADHD classification ROPRNet: Deep learning-assisted recurrence prediction for retinopathy of prematurity Editorial Board
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1