Modal disentangled generative adversarial networks for bidirectional magnetic resonance image synthesis

IF 8 2区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS Engineering Applications of Artificial Intelligence Pub Date : 2025-02-01 Epub Date: 2024-12-10 DOI:10.1016/j.engappai.2024.109817
Liming Xu , Yanrong Lei , Jie Shao , Xianhua Zeng , Weisheng Li
{"title":"Modal disentangled generative adversarial networks for bidirectional magnetic resonance image synthesis","authors":"Liming Xu ,&nbsp;Yanrong Lei ,&nbsp;Jie Shao ,&nbsp;Xianhua Zeng ,&nbsp;Weisheng Li","doi":"10.1016/j.engappai.2024.109817","DOIUrl":null,"url":null,"abstract":"<div><div>Magnetic resonance imaging (MRI) is commonly used both in clinical diagnosis and scientific research. Owing to the high cost, time constraints, and limited application of multi-contrast MRI images obtained from metallic implants, it incurs low throughput and misses a specific modality. Medical image cross-modal synthesis based on Artificial Intelligence (AI) technologies is proposed to synthesize the desired missing modal images. However, it still suffers from low expandability, invisible latent representations, and poor interpretability. We thus propose modal disentanglement generative adversarial networks for bidirectional T1-weighted (T1-w) and T1-weighted (T2-w) medical image synthesis with controllable cross-modal synthesis and disentangled interpretability. Firstly, we construct a cross-modal synthesis model to achieve bidirectional generation between T1-w and T2-w MRI images, which can be easily extended for adaptive modality synthesis without training multiple generators and discriminators. Then, we use an easily trained deep network to disentangle deep representations in latent space and map representations in latent space into pixel space to visualize morphological images and yield multi-contrast MRI images with controllable feature generation. Besides, we construct an easy-to-interpret deep structure by incorporating morphology consistency to preserve edge contours and visualize deep representations in latent space to enable interpretability, which is critical for artificial intelligence oriented to engineering applications and clinical diagnostics. The experiments demonstrate that ours outperforms recent state-of-the-art methods with average improvements of 15.8% structural similarity (SSIM), 12.7% multiscale structural similarity (MSIM), 38.2% peak signal-to-noise ratio (PSNR) and 5.2% visual information fidelity (VIF) on benchmark datasets.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"141 ","pages":"Article 109817"},"PeriodicalIF":8.0000,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197624019766","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/12/10 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Magnetic resonance imaging (MRI) is commonly used both in clinical diagnosis and scientific research. Owing to the high cost, time constraints, and limited application of multi-contrast MRI images obtained from metallic implants, it incurs low throughput and misses a specific modality. Medical image cross-modal synthesis based on Artificial Intelligence (AI) technologies is proposed to synthesize the desired missing modal images. However, it still suffers from low expandability, invisible latent representations, and poor interpretability. We thus propose modal disentanglement generative adversarial networks for bidirectional T1-weighted (T1-w) and T1-weighted (T2-w) medical image synthesis with controllable cross-modal synthesis and disentangled interpretability. Firstly, we construct a cross-modal synthesis model to achieve bidirectional generation between T1-w and T2-w MRI images, which can be easily extended for adaptive modality synthesis without training multiple generators and discriminators. Then, we use an easily trained deep network to disentangle deep representations in latent space and map representations in latent space into pixel space to visualize morphological images and yield multi-contrast MRI images with controllable feature generation. Besides, we construct an easy-to-interpret deep structure by incorporating morphology consistency to preserve edge contours and visualize deep representations in latent space to enable interpretability, which is critical for artificial intelligence oriented to engineering applications and clinical diagnostics. The experiments demonstrate that ours outperforms recent state-of-the-art methods with average improvements of 15.8% structural similarity (SSIM), 12.7% multiscale structural similarity (MSIM), 38.2% peak signal-to-noise ratio (PSNR) and 5.2% visual information fidelity (VIF) on benchmark datasets.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
双向磁共振图像合成的模态解纠缠生成对抗网络
磁共振成像(MRI)在临床诊断和科学研究中都有广泛的应用。由于高成本、时间限制以及从金属植入物获得的多对比MRI图像的应用有限,它导致低通量并且错过了特定的模态。提出了一种基于人工智能(AI)技术的医学图像跨模态合成方法,以合成所需的缺失模态图像。然而,它仍然存在可扩展性低、隐性表示不可见和可解释性差的问题。因此,我们提出了用于双向t1 -加权(T1-w)和t1 -加权(T2-w)医学图像合成的模态解纠缠生成对抗网络,具有可控的跨模态合成和解纠缠解释性。首先,我们构建了一个跨模态合成模型,实现了T1-w和T2-w MRI图像之间的双向生成,该模型可以很容易地扩展到自适应模态合成,而无需训练多个生成器和鉴别器。然后,我们使用一个易于训练的深度网络来解纠缠潜在空间中的深度表征,并将潜在空间中的表征映射到像素空间中,以可视化形态学图像,并生成具有可控特征生成的多对比度MRI图像。此外,我们通过结合形态学一致性来构建易于解释的深层结构以保持边缘轮廓,并在潜在空间中可视化深度表示以实现可解释性,这对于面向工程应用和临床诊断的人工智能至关重要。实验表明,我们的方法优于目前最先进的方法,在基准数据集上平均提高了15.8%的结构相似度(SSIM), 12.7%的多尺度结构相似度(MSIM), 38.2%的峰值信噪比(PSNR)和5.2%的视觉信息保真度(VIF)。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Engineering Applications of Artificial Intelligence
Engineering Applications of Artificial Intelligence 工程技术-工程:电子与电气
CiteScore
9.60
自引率
10.00%
发文量
505
审稿时长
68 days
期刊介绍: Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.
期刊最新文献
Novel artificial intelligence driven model-based control framework for solar–electric vehicle home energy optimization Exploring semantic dependency for reasoning over temporal knowledge graph Enhancing the safety assessment of open-pit mine slopes with interpretable, data-driven stacking learning and three-dimensional stability analysis Dual-domain data enhancement and lightweight deep architecture for robust powder bed defect detection Modeling function-level relationships for vulnerability detection in graph neural networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1