Using a generative adversarial network to generate synthetic MRI images for multi-class automatic segmentation of brain tumors

P. Raut, G. Baldini, M. Schöneck, L. Caldeira
{"title":"Using a generative adversarial network to generate synthetic MRI images for multi-class automatic segmentation of brain tumors","authors":"P. Raut, G. Baldini, M. Schöneck, L. Caldeira","doi":"10.3389/fradi.2023.1336902","DOIUrl":null,"url":null,"abstract":"Challenging tasks such as lesion segmentation, classification, and analysis for the assessment of disease progression can be automatically achieved using deep learning (DL)-based algorithms. DL techniques such as 3D convolutional neural networks are trained using heterogeneous volumetric imaging data such as MRI, CT, and PET, among others. However, DL-based methods are usually only applicable in the presence of the desired number of inputs. In the absence of one of the required inputs, the method cannot be used. By implementing a generative adversarial network (GAN), we aim to apply multi-label automatic segmentation of brain tumors to synthetic images when not all inputs are present. The implemented GAN is based on the Pix2Pix architecture and has been extended to a 3D framework named Pix2PixNIfTI. For this study, 1,251 patients of the BraTS2021 dataset comprising sequences such as T1w, T2w, T1CE, and FLAIR images equipped with respective multi-label segmentation were used. This dataset was used for training the Pix2PixNIfTI model for generating synthetic MRI images of all the image contrasts. The segmentation model, namely DeepMedic, was trained in a five-fold cross-validation manner for brain tumor segmentation and tested using the original inputs as the gold standard. The inference of trained segmentation models was later applied to synthetic images replacing missing input, in combination with other original images to identify the efficacy of generated images in achieving multi-class segmentation. For the multi-class segmentation using synthetic data or lesser inputs, the dice scores were observed to be significantly reduced but remained similar in range for the whole tumor when compared with evaluated original image segmentation (e.g. mean dice of synthetic T2w prediction NC, 0.74 ± 0.30; ED, 0.81 ± 0.15; CET, 0.84 ± 0.21; WT, 0.90 ± 0.08). A standard paired t-tests with multiple comparison correction were performed to assess the difference between all regions (p < 0.05). The study concludes that the use of Pix2PixNIfTI allows us to segment brain tumors when one input image is missing.","PeriodicalId":73101,"journal":{"name":"Frontiers in radiology","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in radiology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/fradi.2023.1336902","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Challenging tasks such as lesion segmentation, classification, and analysis for the assessment of disease progression can be automatically achieved using deep learning (DL)-based algorithms. DL techniques such as 3D convolutional neural networks are trained using heterogeneous volumetric imaging data such as MRI, CT, and PET, among others. However, DL-based methods are usually only applicable in the presence of the desired number of inputs. In the absence of one of the required inputs, the method cannot be used. By implementing a generative adversarial network (GAN), we aim to apply multi-label automatic segmentation of brain tumors to synthetic images when not all inputs are present. The implemented GAN is based on the Pix2Pix architecture and has been extended to a 3D framework named Pix2PixNIfTI. For this study, 1,251 patients of the BraTS2021 dataset comprising sequences such as T1w, T2w, T1CE, and FLAIR images equipped with respective multi-label segmentation were used. This dataset was used for training the Pix2PixNIfTI model for generating synthetic MRI images of all the image contrasts. The segmentation model, namely DeepMedic, was trained in a five-fold cross-validation manner for brain tumor segmentation and tested using the original inputs as the gold standard. The inference of trained segmentation models was later applied to synthetic images replacing missing input, in combination with other original images to identify the efficacy of generated images in achieving multi-class segmentation. For the multi-class segmentation using synthetic data or lesser inputs, the dice scores were observed to be significantly reduced but remained similar in range for the whole tumor when compared with evaluated original image segmentation (e.g. mean dice of synthetic T2w prediction NC, 0.74 ± 0.30; ED, 0.81 ± 0.15; CET, 0.84 ± 0.21; WT, 0.90 ± 0.08). A standard paired t-tests with multiple comparison correction were performed to assess the difference between all regions (p < 0.05). The study concludes that the use of Pix2PixNIfTI allows us to segment brain tumors when one input image is missing.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
使用生成式对抗网络生成合成磁共振成像图像,用于脑肿瘤的多类自动分割
利用基于深度学习(DL)的算法,可以自动完成病变分割、分类和分析等挑战性任务,以评估疾病进展。三维卷积神经网络等深度学习技术是利用 MRI、CT 和 PET 等异构容积成像数据进行训练的。然而,基于 DL 的方法通常只适用于所需输入数量的情况。如果缺少所需的一个输入,该方法就无法使用。通过实施生成对抗网络(GAN),我们的目标是在不存在所有输入的情况下,对合成图像进行多标签脑肿瘤自动分割。所实现的 GAN 基于 Pix2Pix 架构,并已扩展到名为 Pix2PixNIfTI 的三维框架。在这项研究中,使用了 BraTS2021 数据集中的 1,251 名患者,其中包括 T1w、T2w、T1CE 和 FLAIR 图像等序列,并配备了相应的多标签分割。该数据集用于训练 Pix2PixNIfTI 模型,以生成所有图像对比度的合成 MRI 图像。分割模型,即 DeepMedic,以五倍交叉验证的方式进行脑肿瘤分割训练,并使用原始输入作为金标准进行测试。随后,将训练好的分割模型推理应用于替代缺失输入的合成图像,并与其他原始图像相结合,以确定生成图像在实现多类分割方面的功效。使用合成数据或较少的输入进行多类分割时,观察到骰子分数显著降低,但与评估的原始图像分割相比,整个肿瘤的骰子分数范围仍然相似(例如,合成 T2w 预测 NC 的平均骰子分数为 0.74 ± 0.30;ED 为 0.81 ± 0.15;CET 为 0.84 ± 0.21;WT 为 0.90 ± 0.08)。对所有区域之间的差异进行了标准配对 t 检验和多重比较校正(P < 0.05)。研究得出结论,使用 Pix2PixNIfTI 可以在缺少一张输入图像的情况下对脑肿瘤进行分割。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
1.20
自引率
0.00%
发文量
0
期刊最新文献
Wideband radiofrequency pulse sequence for evaluation of myocardial scar in patients with cardiac implantable devices. Value of interventional radiology and their contributions to modern medical systems Feasibility study to unveil the potential: considerations of constrained spherical deconvolution tractography with unsedated neonatal diffusion brain MRI data. Automated intracranial vessel segmentation of 4D flow MRI data in patients with atherosclerotic stenosis using a convolutional neural network Standardized evaluation of the extent of resection in glioblastoma with automated early post-operative segmentation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1