Multiscale segmentation net for segregating heterogeneous brain tumors: Gliomas on multimodal MR images

IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Image and Vision Computing Pub Date : 2024-07-18 DOI:10.1016/j.imavis.2024.105191
{"title":"Multiscale segmentation net for segregating heterogeneous brain tumors: Gliomas on multimodal MR images","authors":"","doi":"10.1016/j.imavis.2024.105191","DOIUrl":null,"url":null,"abstract":"<div><p>In this research, the 3D volumetric segmentation of heterogeneous brain tumors such as Gliomas- anaplastic astrocytoma, and Glioblastoma Multiforme (GBM) is performed to extract enhancing tumor (ET), whole tumor (WT), and tumor core (TC) regions using T1, T2, and FLAIR images. Therefore, a deep learning-based encoder-decoder architecture named “MS-SegNet” using 3D multi-scale convolutional layers is proposed. The proposed architecture employs multi-scale feature extraction (MS-FE) block the filter size 3 × 3 × 3 to extract confined information like tumor boundary and edges of necrotic part. The filter of size 5 × 5 × 5 focuses on varied features like shape, size, and location of tumor region with edema. The local and global features from different MR modalities are extracted for segmenting thin and meshed boundaries of heterogeneous tumors between anatomical sub-regions like peritumoral edema, enhancing tumor, and necrotic tumor core. The learning parameters on introducing the MS-FE block are reduced to 10 million, which is much less than other architectures like 3D-Unet which takes into consideration 27 million features leading to the consumption of less computational power. A customized loss function is also prophesied based on a combination of dice loss and focal loss along with metrics such as accuracy and Intersection over Union (IoU) i.e. the overlapping of ground truth mask and predicted value for addressing the class imbalance problem. For evaluating the efficacy of the proposed method, four evaluation metrics such as Dice Coefficient (DSC), Sensitivity, Specificity, and Hausdorff95 distance (H95) are employed for analyzing the model's overall performance. It is observed that the proposed MS-SegNet architecture achieved the DSC of 0.81, 0.91, and 0.83 on BraTS 2020; 0.86, 0.92, and 0.84 on BraTS 2021 for ET, WT, and TC respectively. The developed model is also tested on a real-time dataset collected from the Post Graduate Institute of Medical Education &amp; Research (PGIMER), Chandigarh. The DSC of 0.79, 0.76, and 0.68 for ET, WT, and TC respectively on the real-time dataset. These findings show that deep learning models with enhanced feature extraction capabilities can be readily trained to attain high accuracy in segmenting heterogeneous brain tumors and hold promising results. In the future, other tumor datasets will be explored for the detection and treatment planning of brain tumors to check the effectiveness of the model in real-world healthcare environments.</p></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":null,"pages":null},"PeriodicalIF":4.2000,"publicationDate":"2024-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Image and Vision Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0262885624002968","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

In this research, the 3D volumetric segmentation of heterogeneous brain tumors such as Gliomas- anaplastic astrocytoma, and Glioblastoma Multiforme (GBM) is performed to extract enhancing tumor (ET), whole tumor (WT), and tumor core (TC) regions using T1, T2, and FLAIR images. Therefore, a deep learning-based encoder-decoder architecture named “MS-SegNet” using 3D multi-scale convolutional layers is proposed. The proposed architecture employs multi-scale feature extraction (MS-FE) block the filter size 3 × 3 × 3 to extract confined information like tumor boundary and edges of necrotic part. The filter of size 5 × 5 × 5 focuses on varied features like shape, size, and location of tumor region with edema. The local and global features from different MR modalities are extracted for segmenting thin and meshed boundaries of heterogeneous tumors between anatomical sub-regions like peritumoral edema, enhancing tumor, and necrotic tumor core. The learning parameters on introducing the MS-FE block are reduced to 10 million, which is much less than other architectures like 3D-Unet which takes into consideration 27 million features leading to the consumption of less computational power. A customized loss function is also prophesied based on a combination of dice loss and focal loss along with metrics such as accuracy and Intersection over Union (IoU) i.e. the overlapping of ground truth mask and predicted value for addressing the class imbalance problem. For evaluating the efficacy of the proposed method, four evaluation metrics such as Dice Coefficient (DSC), Sensitivity, Specificity, and Hausdorff95 distance (H95) are employed for analyzing the model's overall performance. It is observed that the proposed MS-SegNet architecture achieved the DSC of 0.81, 0.91, and 0.83 on BraTS 2020; 0.86, 0.92, and 0.84 on BraTS 2021 for ET, WT, and TC respectively. The developed model is also tested on a real-time dataset collected from the Post Graduate Institute of Medical Education & Research (PGIMER), Chandigarh. The DSC of 0.79, 0.76, and 0.68 for ET, WT, and TC respectively on the real-time dataset. These findings show that deep learning models with enhanced feature extraction capabilities can be readily trained to attain high accuracy in segmenting heterogeneous brain tumors and hold promising results. In the future, other tumor datasets will be explored for the detection and treatment planning of brain tumors to check the effectiveness of the model in real-world healthcare environments.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
用于分离异质脑肿瘤的多尺度分割网--神经胶质瘤多模态磁共振图像
本研究利用 T1、T2 和 FLAIR 图像对胶质瘤-无弹性星形细胞瘤和多形性胶质母细胞瘤(GBM)等异质脑肿瘤进行三维容积分割,以提取增强肿瘤(ET)、全瘤(WT)和瘤核(TC)区域。因此,我们提出了一种基于深度学习的编码器-解码器架构,名为 "MS-SegNet",使用三维多尺度卷积层。所提出的架构采用多尺度特征提取(MS-FE)块,滤波器大小为 3 × 3 × 3,以提取肿瘤边界和坏死部分边缘等密闭信息。尺寸为 5 × 5 × 5 的滤波器侧重于各种特征,如肿瘤区域的形状、大小和水肿位置。从不同的磁共振成像模式中提取局部和全局特征,用于分割异质肿瘤在瘤周水肿、增强肿瘤和坏死瘤核等解剖亚区域之间的细边界和网状边界。引入 MS-FE 模块后,学习参数减少到 1 千万个,远远少于其他架构,如考虑到 2,700 万个特征的 3D-Unet 架构,从而降低了计算能力的消耗。为解决类别不平衡问题,还根据骰子损失和焦点损失的组合以及准确度和交集大于联合(IoU)(即地面实况掩码和预测值的重叠)等指标预言了一个定制的损失函数。为了评估所提出方法的功效,采用了四种评估指标,如骰子系数(DSC)、灵敏度、特异度和豪斯多夫 95 距离(H95),以分析模型的整体性能。结果表明,所提出的 MS-SegNet 架构在 BraTS 2020 上的 DSC 分别为 0.81、0.91 和 0.83;在 BraTS 2021 上,ET、WT 和 TC 的 DSC 分别为 0.86、0.92 和 0.84。开发的模型还在昌迪加尔医学教育与研究研究生院(PGIMER)收集的实时数据集上进行了测试。在实时数据集上,ET、WT 和 TC 的 DSC 分别为 0.79、0.76 和 0.68。这些研究结果表明,具有增强的特征提取能力的深度学习模型可以很容易地进行训练,从而在分割异质脑肿瘤时达到较高的准确率,并取得了可喜的成果。未来,还将探索其他肿瘤数据集,用于脑肿瘤的检测和治疗规划,以检验模型在真实医疗环境中的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Image and Vision Computing
Image and Vision Computing 工程技术-工程:电子与电气
CiteScore
8.50
自引率
8.50%
发文量
143
审稿时长
7.8 months
期刊介绍: Image and Vision Computing has as a primary aim the provision of an effective medium of interchange for the results of high quality theoretical and applied research fundamental to all aspects of image interpretation and computer vision. The journal publishes work that proposes new image interpretation and computer vision methodology or addresses the application of such methods to real world scenes. It seeks to strengthen a deeper understanding in the discipline by encouraging the quantitative comparison and performance evaluation of the proposed methodology. The coverage includes: image interpretation, scene modelling, object recognition and tracking, shape analysis, monitoring and surveillance, active vision and robotic systems, SLAM, biologically-inspired computer vision, motion analysis, stereo vision, document image understanding, character and handwritten text recognition, face and gesture recognition, biometrics, vision-based human-computer interaction, human activity and behavior understanding, data fusion from multiple sensor inputs, image databases.
期刊最新文献
A dictionary learning based unsupervised neural network for single image compressed sensing Unbiased scene graph generation via head-tail cooperative network with self-supervised learning UIR-ES: An unsupervised underwater image restoration framework with equivariance and stein unbiased risk estimator A new deepfake detection model for responding to perception attacks in embodied artificial intelligence Ground4Act: Leveraging visual-language model for collaborative pushing and grasping in clutter
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1