Self-supervised U-transformer network with mask reconstruction for metal artifact reduction.

IF 3.4 3区 医学 Q2 ENGINEERING, BIOMEDICAL Physics in medicine and biology Pub Date : 2025-03-10 DOI:10.1088/1361-6560/adbaae
Fanning Kong, Zaifeng Shi, Huaisheng Cao, Yudong Hao, Qingjie Cao
{"title":"Self-supervised U-transformer network with mask reconstruction for metal artifact reduction.","authors":"Fanning Kong, Zaifeng Shi, Huaisheng Cao, Yudong Hao, Qingjie Cao","doi":"10.1088/1361-6560/adbaae","DOIUrl":null,"url":null,"abstract":"<p><p><i>Objective</i>. Metal artifacts severely damaged human tissue information from the computed tomography (CT) image, posing significant challenges to disease diagnosis. Deep learning has been widely explored for the metal artifact reduction (MAR) task. Nevertheless, paired metal artifact CT datasets suitable for training do not exist in reality. Although the synthetic CT image dataset provides additional training data, the trained networks still generalize poorly to real metal artifact data.<i>Approach.</i>A self-supervised U-shaped transformer network is proposed to focus on model generalizability enhancement in MAR tasks. This framework consists of a self-supervised mask reconstruction pre-text task and a down-stream task. In the pre-text task, the CT images are randomly corrupted by masks. They are recovered with themselves as the label, aiming at acquiring the artifacts and tissue structure of the actual physical situation. Down-stream task fine-tunes MAR target through labeled images. Utilizing the multi-layer long-range feature extraction capabilities of the Transformer efficiently captures features of metal artifacts. The incorporation of the MAR bottleneck allows for the distinction of metal artifact features through cross-channel self-attention.<i>Main result</i>. Experiments demonstrate that the framework maintains strong generalization ability in the MAR task, effectively preserving tissue details while suppressing metal artifacts. The results achieved a peak signal-to-noise ratio of 43.86 dB and a structural similarity index of 0.9863 while ensuring the efficiency of the model inference. In addition, the Dice coefficient and mean intersection over union are improved by 11.70% and 9.51% in the segmentation of the MAR image, respectively.<i>Significance.</i>The combination of unlabeled real-artifact CT images and labeled synthetic-artifact CT images facilitates a self-supervised learning process that positively contributes to model generalizability.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4000,"publicationDate":"2025-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Physics in medicine and biology","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1088/1361-6560/adbaae","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0

Abstract

Objective. Metal artifacts severely damaged human tissue information from the computed tomography (CT) image, posing significant challenges to disease diagnosis. Deep learning has been widely explored for the metal artifact reduction (MAR) task. Nevertheless, paired metal artifact CT datasets suitable for training do not exist in reality. Although the synthetic CT image dataset provides additional training data, the trained networks still generalize poorly to real metal artifact data.Approach.A self-supervised U-shaped transformer network is proposed to focus on model generalizability enhancement in MAR tasks. This framework consists of a self-supervised mask reconstruction pre-text task and a down-stream task. In the pre-text task, the CT images are randomly corrupted by masks. They are recovered with themselves as the label, aiming at acquiring the artifacts and tissue structure of the actual physical situation. Down-stream task fine-tunes MAR target through labeled images. Utilizing the multi-layer long-range feature extraction capabilities of the Transformer efficiently captures features of metal artifacts. The incorporation of the MAR bottleneck allows for the distinction of metal artifact features through cross-channel self-attention.Main result. Experiments demonstrate that the framework maintains strong generalization ability in the MAR task, effectively preserving tissue details while suppressing metal artifacts. The results achieved a peak signal-to-noise ratio of 43.86 dB and a structural similarity index of 0.9863 while ensuring the efficiency of the model inference. In addition, the Dice coefficient and mean intersection over union are improved by 11.70% and 9.51% in the segmentation of the MAR image, respectively.Significance.The combination of unlabeled real-artifact CT images and labeled synthetic-artifact CT images facilitates a self-supervised learning process that positively contributes to model generalizability.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于掩模重构的自监督u型变压器网络。
目的:金属伪影严重破坏了CT图像中的人体组织信息,对疾病诊断提出了重大挑战。深度学习(DL)在金属伪影还原(MAR)任务中得到了广泛的探索。然而,现实中并不存在适合训练的配对金属伪影CT数据集。尽管合成CT图像数据集提供了额外的训练数据,但训练后的网络对真实金属伪影数据的泛化能力仍然很差。 ;提出了一种自监督u型变压器网络(SUTransNet),以增强MAR任务中的模型泛化能力。该框架由自监督掩码重构文本前任务和下游任务组成。在pre-text任务中,CT图像被蒙版随机破坏。它们以自身为标签进行复原,目的是获取实际物理状态的人工制品和组织结构。下游任务通过标记图像对MAR目标进行微调。利用Transformer的多层远程特征提取功能,可以有效地捕获金属工件的特征。MAR瓶颈的结合允许通过跨通道自关注来区分金属工件特征。实验表明,该框架在MAR任务中保持了较强的泛化能力,有效地保留了组织细节,同时抑制了金属伪影。在保证模型推理效率的前提下,峰值信噪比(PSNR)为43.86 dB,结构相似性指数(SSIM)为0.9863。此外,在分割MAR图像时,Dice系数和Mean Intersection over Union (MIoU)分别提高了11.70%和9.51% . ;将未标记的真实伪影CT图像与标记的合成伪影CT图像相结合,促进了自监督学习过程,对模型的泛化有积极的贡献。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Physics in medicine and biology
Physics in medicine and biology 医学-工程:生物医学
CiteScore
6.50
自引率
14.30%
发文量
409
审稿时长
2 months
期刊介绍: The development and application of theoretical, computational and experimental physics to medicine, physiology and biology. Topics covered are: therapy physics (including ionizing and non-ionizing radiation); biomedical imaging (e.g. x-ray, magnetic resonance, ultrasound, optical and nuclear imaging); image-guided interventions; image reconstruction and analysis (including kinetic modelling); artificial intelligence in biomedical physics and analysis; nanoparticles in imaging and therapy; radiobiology; radiation protection and patient dose monitoring; radiation dosimetry
期刊最新文献
Evaluation of radiological properties and anisotropy with air channels analysis in 3D-printed flexible lung-mimicking materials for radiotherapy. Combining proton FLASH and spatially fractionated radiotherapy: experimental and simulation-based dosimetric characterization. Transformer-encoded nnU-Net with local region perceptron and contrastive learning (TLC-nnUNet) for multiple brain metastasis detection and delineation. Functional-based multi-omics early prediction of radiation pneumonitis in NSCLC using AI-generated perfusion and ventilation from planning CT. Estimating time-dependent cell survival in particle beam thermoradiotherapy with the dynamic temperature-dependent stochastic microdosimetric kinetic model.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1