CMAF-Net: a cross-modal attention fusion-based deep neural network for incomplete multi-modal brain tumor segmentation.

IF 2.9 2区 医学 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Quantitative Imaging in Medicine and Surgery Pub Date : 2024-07-01 Epub Date: 2024-06-27 DOI:10.21037/qims-24-9
Kangkang Sun, Jiangyi Ding, Qixuan Li, Wei Chen, Heng Zhang, Jiawei Sun, Zhuqing Jiao, Xinye Ni
{"title":"CMAF-Net: a cross-modal attention fusion-based deep neural network for incomplete multi-modal brain tumor segmentation.","authors":"Kangkang Sun, Jiangyi Ding, Qixuan Li, Wei Chen, Heng Zhang, Jiawei Sun, Zhuqing Jiao, Xinye Ni","doi":"10.21037/qims-24-9","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>The information between multimodal magnetic resonance imaging (MRI) is complementary. Combining multiple modalities for brain tumor image segmentation can improve segmentation accuracy, which has great significance for disease diagnosis and treatment. However, different degrees of missing modality data often occur in clinical practice, which may lead to serious performance degradation or even failure of brain tumor segmentation methods relying on full-modality sequences to complete the segmentation task. To solve the above problems, this study aimed to design a new deep learning network for incomplete multimodal brain tumor segmentation.</p><p><strong>Methods: </strong>We propose a novel cross-modal attention fusion-based deep neural network (CMAF-Net) for incomplete multimodal brain tumor segmentation, which is based on a three-dimensional (3D) U-Net architecture with encoding and decoding structure, a 3D Swin block, and a cross-modal attention fusion (CMAF) block. A convolutional encoder is initially used to extract the specific features from different modalities, and an effective 3D Swin block is constructed to model the long-range dependencies to obtain richer information for brain tumor segmentation. Then, a cross-attention based CMAF module is proposed that can deal with different missing modality situations by fusing features between different modalities to learn the shared representations of the tumor regions. Finally, the fused latent representation is decoded to obtain the final segmentation result. Additionally, channel attention module (CAM) and spatial attention module (SAM) are incorporated into the network to further improve the robustness of the model; the CAM to help focus on important feature channels, and the SAM to learn the importance of different spatial regions.</p><p><strong>Results: </strong>Evaluation experiments on the widely-used BraTS 2018 and BraTS 2020 datasets demonstrated the effectiveness of the proposed CMAF-Net which achieved average Dice scores of 87.9%, 81.8%, and 64.3%, as well as Hausdorff distances of 4.21, 5.35, and 4.02 for whole tumor, tumor core, and enhancing tumor on the BraTS 2020 dataset, respectively, outperforming several state-of-the-art segmentation methods in missing modalities situations.</p><p><strong>Conclusions: </strong>The experimental results show that the proposed CMAF-Net can achieve accurate brain tumor segmentation in the case of missing modalities with promising application potential.</p>","PeriodicalId":54267,"journal":{"name":"Quantitative Imaging in Medicine and Surgery","volume":null,"pages":null},"PeriodicalIF":2.9000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11250309/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Quantitative Imaging in Medicine and Surgery","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.21037/qims-24-9","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/6/27 0:00:00","PubModel":"Epub","JCR":"Q2","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0

Abstract

Background: The information between multimodal magnetic resonance imaging (MRI) is complementary. Combining multiple modalities for brain tumor image segmentation can improve segmentation accuracy, which has great significance for disease diagnosis and treatment. However, different degrees of missing modality data often occur in clinical practice, which may lead to serious performance degradation or even failure of brain tumor segmentation methods relying on full-modality sequences to complete the segmentation task. To solve the above problems, this study aimed to design a new deep learning network for incomplete multimodal brain tumor segmentation.

Methods: We propose a novel cross-modal attention fusion-based deep neural network (CMAF-Net) for incomplete multimodal brain tumor segmentation, which is based on a three-dimensional (3D) U-Net architecture with encoding and decoding structure, a 3D Swin block, and a cross-modal attention fusion (CMAF) block. A convolutional encoder is initially used to extract the specific features from different modalities, and an effective 3D Swin block is constructed to model the long-range dependencies to obtain richer information for brain tumor segmentation. Then, a cross-attention based CMAF module is proposed that can deal with different missing modality situations by fusing features between different modalities to learn the shared representations of the tumor regions. Finally, the fused latent representation is decoded to obtain the final segmentation result. Additionally, channel attention module (CAM) and spatial attention module (SAM) are incorporated into the network to further improve the robustness of the model; the CAM to help focus on important feature channels, and the SAM to learn the importance of different spatial regions.

Results: Evaluation experiments on the widely-used BraTS 2018 and BraTS 2020 datasets demonstrated the effectiveness of the proposed CMAF-Net which achieved average Dice scores of 87.9%, 81.8%, and 64.3%, as well as Hausdorff distances of 4.21, 5.35, and 4.02 for whole tumor, tumor core, and enhancing tumor on the BraTS 2020 dataset, respectively, outperforming several state-of-the-art segmentation methods in missing modalities situations.

Conclusions: The experimental results show that the proposed CMAF-Net can achieve accurate brain tumor segmentation in the case of missing modalities with promising application potential.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
CMAF-Net:基于跨模态注意力融合的深度神经网络,用于不完整的多模态脑肿瘤分割。
背景:多模态磁共振成像(MRI)之间的信息是互补的。结合多种模态进行脑肿瘤图像分割可以提高分割的准确性,对疾病的诊断和治疗具有重要意义。然而,在临床实践中经常会出现不同程度的模态数据缺失,这可能会导致依赖全模态序列完成分割任务的脑肿瘤分割方法性能严重下降甚至失效。为解决上述问题,本研究旨在设计一种新的深度学习网络,用于不完全多模态脑肿瘤分割:我们提出了一种用于不完全多模态脑肿瘤分割的新型基于跨模态注意力融合的深度神经网络(CMAF-Net),它基于具有编码和解码结构的三维(3D)U-Net 架构、三维 Swin 块和跨模态注意力融合(CMAF)块。首先使用卷积编码器从不同模态中提取特定特征,然后构建有效的三维 Swin 模块来模拟长程依赖关系,从而获得更丰富的脑肿瘤分割信息。然后,提出一种基于交叉关注的 CMAF 模块,通过融合不同模态之间的特征来学习肿瘤区域的共享表征,从而处理不同的模态缺失情况。最后,对融合后的潜在表征进行解码,得到最终的分割结果。此外,为了进一步提高模型的鲁棒性,还在网络中加入了通道关注模块(CAM)和空间关注模块(SAM);CAM 用于帮助关注重要的特征通道,SAM 用于学习不同空间区域的重要性:在广泛使用的 BraTS 2018 和 BraTS 2020 数据集上进行的评估实验证明了所提出的 CMAF 网络的有效性,在 BraTS 2020 数据集上,整个肿瘤、肿瘤核心和增强肿瘤的平均 Dice 分数分别为 87.9%、81.8% 和 64.3%,Hausdorff 距离分别为 4.21、5.35 和 4.02,在缺失模态情况下优于几种最先进的分割方法:实验结果表明,所提出的 CMAF-Net 可以在模态缺失的情况下实现精确的脑肿瘤分割,具有广阔的应用前景。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Quantitative Imaging in Medicine and Surgery
Quantitative Imaging in Medicine and Surgery Medicine-Radiology, Nuclear Medicine and Imaging
CiteScore
4.20
自引率
17.90%
发文量
252
期刊介绍: Information not localized
期刊最新文献
Deep learning-based quantitative morphological study of anteroposterior digital radiographs of the lumbar spine. Automatic measurement of anatomical parameters of the lumbar vertebral body and the intervertebral disc on radiographs by deep learning. Bone age assessment by multi-granularity and multi-attention feature encoding. Can cereal-based oral contrast agents-assisted ultrasound become an alternative to non-contrast magnetic resonance imaging (MRI) in radiological follow-up for pancreatic cystic lesions? Cardiac intravoxel incoherent motion diffusion-weighted imaging to assess myocardial microcirculation dysfunction in hypertension.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1