Joint segmentation of tumors in 3D PET-CT images with a network fusing multi-view and multi-modal information.

IF 3.3 3区 医学 Q2 ENGINEERING, BIOMEDICAL Physics in medicine and biology Pub Date : 2024-10-07 DOI:10.1088/1361-6560/ad7f1b
HaoYang Zheng, Wei Zou, Nan Hu, Jiajun Wang
{"title":"Joint segmentation of tumors in 3D PET-CT images with a network fusing multi-view and multi-modal information.","authors":"HaoYang Zheng, Wei Zou, Nan Hu, Jiajun Wang","doi":"10.1088/1361-6560/ad7f1b","DOIUrl":null,"url":null,"abstract":"<p><p><i>Objective</i>. Joint segmentation of tumors in positron emission tomography-computed tomography (PET-CT) images is crucial for precise treatment planning. However, current segmentation methods often use addition or concatenation to fuse PET and CT images, which potentially overlooks the nuanced interplay between these modalities. Additionally, these methods often neglect multi-view information that is helpful for more accurately locating and segmenting the target structure. This study aims to address these disadvantages and develop a deep learning-based algorithm for joint segmentation of tumors in PET-CT images.<i>Approach</i>. To address these limitations, we propose the Multi-view Information Enhancement and Multi-modal Feature Fusion Network (MIEMFF-Net) for joint tumor segmentation in three-dimensional PET-CT images. Our model incorporates a dynamic multi-modal fusion strategy to effectively exploit the metabolic and anatomical information from PET and CT images and a multi-view information enhancement strategy to effectively recover the lost information during upsamping. A Multi-scale Spatial Perception Block is proposed to effectively extract information from different views and reduce redundancy interference in the multi-view feature extraction process.<i>Main results</i>. The proposed MIEMFF-Net achieved a Dice score of 83.93%, a Precision of 81.49%, a Sensitivity of 87.89% and an IOU of 69.27% on the Soft Tissue Sarcomas dataset and a Dice score of 76.83%, a Precision of 86.21%, a Sensitivity of 80.73% and an IOU of 65.15% on the AutoPET dataset.<i>Significance</i>. Experimental results demonstrate that MIEMFF-Net outperforms existing state-of-the-art models which implies potential applications of the proposed method in clinical practice.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.3000,"publicationDate":"2024-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Physics in medicine and biology","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1088/1361-6560/ad7f1b","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0

Abstract

Objective. Joint segmentation of tumors in positron emission tomography-computed tomography (PET-CT) images is crucial for precise treatment planning. However, current segmentation methods often use addition or concatenation to fuse PET and CT images, which potentially overlooks the nuanced interplay between these modalities. Additionally, these methods often neglect multi-view information that is helpful for more accurately locating and segmenting the target structure. This study aims to address these disadvantages and develop a deep learning-based algorithm for joint segmentation of tumors in PET-CT images.Approach. To address these limitations, we propose the Multi-view Information Enhancement and Multi-modal Feature Fusion Network (MIEMFF-Net) for joint tumor segmentation in three-dimensional PET-CT images. Our model incorporates a dynamic multi-modal fusion strategy to effectively exploit the metabolic and anatomical information from PET and CT images and a multi-view information enhancement strategy to effectively recover the lost information during upsamping. A Multi-scale Spatial Perception Block is proposed to effectively extract information from different views and reduce redundancy interference in the multi-view feature extraction process.Main results. The proposed MIEMFF-Net achieved a Dice score of 83.93%, a Precision of 81.49%, a Sensitivity of 87.89% and an IOU of 69.27% on the Soft Tissue Sarcomas dataset and a Dice score of 76.83%, a Precision of 86.21%, a Sensitivity of 80.73% and an IOU of 65.15% on the AutoPET dataset.Significance. Experimental results demonstrate that MIEMFF-Net outperforms existing state-of-the-art models which implies potential applications of the proposed method in clinical practice.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
利用融合多视角和多模态信息的网络对三维 PET-CT 图像中的肿瘤进行联合分割。
目的:PET-CT 图像中肿瘤的联合分割对于精确的治疗计划至关重要。然而,目前的分割方法通常使用加法或并法来融合 PET 和 CT 图像,这可能会忽略这些模式之间微妙的相互作用。此外,这些方法往往忽略了多视角信息,而这些信息有助于更准确地定位和分割目标结构。本研究旨在解决这些缺点,并开发一种基于深度学习的算法,用于 PET-CT 图像中的肿瘤联合分割。针对这些局限性,我们提出了多视图信息增强和多模态特征融合网络(MIEMFF-Net),用于三维 PET-CT 图像中的联合肿瘤分割。我们的模型融合了动态多模态融合策略和多视图信息增强策略,前者可有效利用 PET 和 CT 图像中的代谢和解剖信息,后者可有效恢复上采样过程中丢失的信息。提出了多尺度空间感知块,以有效提取不同视图的信息,减少多视图特征提取过程中的冗余干扰。提出的 MIEMFF-Net 在 STS 数据集上的 Dice 得分为 83.93%,精确度为 81.49%,灵敏度为 87.89%,IOU 为 69.27%;在 AutoPET 数据集上的 Dice 得分为 76.83%,精确度为 86.21%,灵敏度为 80.73%,IOU 为 65.15%。实验结果表明,MIEMFF-Net 优于现有的最先进(SOTA)模型,这意味着所提出的方法有可能应用于临床实践。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Physics in medicine and biology
Physics in medicine and biology 医学-工程:生物医学
CiteScore
6.50
自引率
14.30%
发文量
409
审稿时长
2 months
期刊介绍: The development and application of theoretical, computational and experimental physics to medicine, physiology and biology. Topics covered are: therapy physics (including ionizing and non-ionizing radiation); biomedical imaging (e.g. x-ray, magnetic resonance, ultrasound, optical and nuclear imaging); image-guided interventions; image reconstruction and analysis (including kinetic modelling); artificial intelligence in biomedical physics and analysis; nanoparticles in imaging and therapy; radiobiology; radiation protection and patient dose monitoring; radiation dosimetry
期刊最新文献
Deep learning methods for 3D magnetic resonance image denoising, bias field and motion artifact correction: a comprehensive review. Effects of spot size errors in DynamicARC pencil beam scanning proton therapy planning. Verification of linear energy transfer optimized carbon-ion radiotherapy. Automated planning of curved needle channels in 3D printed patient-tailored applicators for cervical cancer brachytherapy. Comparison of contrast-enhanced ultrasound imaging (CEUS) and super-resolution ultrasound (SRU) for the quantification of ischaemia flow redistribution: a theoretical study.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1