首页 > 最新文献

IEEE transactions on medical imaging最新文献

英文 中文
A Chain of Diagnosis Framework for Accurate and Explainable Radiology Report Generation. 生成准确和可解释的放射学报告的诊断链框架。
Pub Date : 2025-12-01 DOI: 10.1109/TMI.2025.3585765
Haibo Jin, Haoxuan Che, Sunan He, Hao Chen

Despite the progress of radiology report generation (RRG), existing works face two challenges: 1) The performances in clinical efficacy are unsatisfactory, especially for lesion attributes description; 2) the generated text lacks explainability, making it difficult for radiologists to trust the results. To address the challenges, we focus on a trustworthy RRG model, which not only generates accurate descriptions of abnormalities, but also provides basis of its predictions. To this end, we propose a framework named chain of diagnosis (CoD), which maintains a chain of diagnostic process for clinically accurate and explainable RRG. It first generates question-answer (QA) pairs via diagnostic conversation to extract key findings, then prompts a large language model with QA diagnoses for accurate generation. To enhance explainability, a diagnosis grounding module is designed to match QA diagnoses and generated sentences, where the diagnoses act as a reference. Moreover, a lesion grounding module is designed to locate abnormalities in the image, further improving the working efficiency of radiologists. To facilitate label-efficient training, we propose an omni-supervised learning strategy with clinical consistency to leverage various types of annotations from different datasets. Our efforts lead to 1) an omni-labeled RRG dataset with QA pairs and lesion boxes; 2) a evaluation tool for assessing the accuracy of reports in describing lesion location and severity; 3) extensive experiments to demonstrate the effectiveness of CoD, where it outperforms both specialist and generalist models consistently on two RRG benchmarks and shows promising explainability by accurately grounding generated sentences to QA diagnoses and images.

尽管放射学报告生成(RRG)取得了进展,但现有工作面临两个挑战:1)临床疗效表现不理想,特别是对病变属性的描述;2)生成的文本缺乏可解释性,使放射科医生难以信任结果。为了应对这些挑战,我们将重点放在一个值得信赖的RRG模型上,该模型不仅可以生成对异常的准确描述,还可以为其预测提供基础。为此,我们提出了一个名为诊断链(CoD)的框架,该框架维持了临床准确和可解释的RRG的诊断过程链。它首先通过诊断对话生成问答(QA)对,以提取关键发现,然后用QA诊断提示一个大型语言模型,以准确生成。为了提高可解释性,设计了诊断基础模块来匹配QA诊断和生成的句子,其中诊断作为参考。同时设计病灶接地模块,定位图像中的异常,进一步提高放射科医生的工作效率。为了促进标签高效训练,我们提出了一种具有临床一致性的全监督学习策略,以利用来自不同数据集的各种类型的注释。我们的努力导致1)一个带有QA对和病变盒的全标记RRG数据集;2)评估报告描述病变位置和严重程度的准确性的评估工具;3)广泛的实验来证明CoD的有效性,在两个RRG基准上,它始终优于专家和通才模型,并通过准确地将生成的句子与QA诊断和图像相结合,显示出有希望的可解释性。
{"title":"A Chain of Diagnosis Framework for Accurate and Explainable Radiology Report Generation.","authors":"Haibo Jin, Haoxuan Che, Sunan He, Hao Chen","doi":"10.1109/TMI.2025.3585765","DOIUrl":"10.1109/TMI.2025.3585765","url":null,"abstract":"<p><p>Despite the progress of radiology report generation (RRG), existing works face two challenges: 1) The performances in clinical efficacy are unsatisfactory, especially for lesion attributes description; 2) the generated text lacks explainability, making it difficult for radiologists to trust the results. To address the challenges, we focus on a trustworthy RRG model, which not only generates accurate descriptions of abnormalities, but also provides basis of its predictions. To this end, we propose a framework named chain of diagnosis (CoD), which maintains a chain of diagnostic process for clinically accurate and explainable RRG. It first generates question-answer (QA) pairs via diagnostic conversation to extract key findings, then prompts a large language model with QA diagnoses for accurate generation. To enhance explainability, a diagnosis grounding module is designed to match QA diagnoses and generated sentences, where the diagnoses act as a reference. Moreover, a lesion grounding module is designed to locate abnormalities in the image, further improving the working efficiency of radiologists. To facilitate label-efficient training, we propose an omni-supervised learning strategy with clinical consistency to leverage various types of annotations from different datasets. Our efforts lead to 1) an omni-labeled RRG dataset with QA pairs and lesion boxes; 2) a evaluation tool for assessing the accuracy of reports in describing lesion location and severity; 3) extensive experiments to demonstrate the effectiveness of CoD, where it outperforms both specialist and generalist models consistently on two RRG benchmarks and shows promising explainability by accurately grounding generated sentences to QA diagnoses and images.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":"4986-4997"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144562423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mutualistic Multi-Network Noisy Label Learning (MMNNLL) Method and Its Application to Transdiagnostic Classification of Bipolar Disorder and Schizophrenia. 互惠多网络噪声标签学习(MMNNLL)方法及其在双相情感障碍和精神分裂症跨诊断分类中的应用。
Pub Date : 2025-12-01 DOI: 10.1109/TMI.2025.3585880
Yuhui Du, Zheng Wang, Ju Niu, Yulong Wang, Godfrey D Pearlson, Vince D Calhoun

The subjective nature of diagnosing mental disorders complicates achieving accurate diagnoses. The complex relationship among disorders further exacerbates this issue, particularly in clinical practice where conditions like bipolar disorder (BP) and schizophrenia (SZ) can present similar clinical symptoms and cognitive impairments. To address these challenges, this paper proposes a mutualistic multi-network noisy label learning (MMNNLL) method, which aims to enhance diagnostic accuracy by leveraging neuroimaging data under the presence of potential clinical diagnosis bias or errors. MMNNLL effectively utilizes multiple deep neural networks (DNNs) for learning from data with noisy labels by maximizing the consistency among DNNs in identifying and utilizing samples with clean and noisy labels. Experimental results on public CIFAR-10 and PathMNIST datasets demonstrate the effectiveness of our method in classifying independent test data across various types and levels of label noise. Additionally, our MMNNLL method significantly outperforms state-of-the-art noisy label learning methods. When applied to brain functional connectivity data from BP and SZ patients, our method identifies two biotypes that show more pronounced group differences, and improved classification accuracy compared to the original clinical categories, using both traditional machine learning and advanced deep learning techniques. In summary, our method effectively addresses the possible inaccuracy in nosology of mental disorders and achieves transdiagnostic classification through robust noisy label learning via multi-network collaboration and competition.

诊断精神障碍的主观性使准确诊断变得复杂。疾病之间的复杂关系进一步加剧了这一问题,特别是在临床实践中,双相情感障碍(BP)和精神分裂症(SZ)等疾病可以表现出类似的临床症状和认知障碍。为了解决这些挑战,本文提出了一种互惠的多网络噪声标签学习(MMNNLL)方法,该方法旨在利用存在潜在临床诊断偏差或错误的神经影像学数据来提高诊断准确性。MMNNLL有效地利用多个深度神经网络(dnn)从带有噪声标签的数据中学习,通过最大化dnn在识别和利用带有干净和噪声标签的样本时的一致性。在公开的CIFAR-10和PathMNIST数据集上的实验结果表明,我们的方法在对不同类型和级别的标签噪声的独立测试数据进行分类方面是有效的。此外,我们的MMNNLL方法显著优于最先进的噪声标签学习方法。当应用于BP和SZ患者的脑功能连接数据时,我们的方法识别出两种生物型,与原始临床分类相比,它们表现出更明显的组差异,并且使用传统的机器学习和先进的深度学习技术,提高了分类精度。综上所述,我们的方法有效地解决了精神障碍分类学中可能存在的不准确性,并通过多网络协作和竞争,通过鲁棒噪声标签学习实现了跨诊断分类。
{"title":"Mutualistic Multi-Network Noisy Label Learning (MMNNLL) Method and Its Application to Transdiagnostic Classification of Bipolar Disorder and Schizophrenia.","authors":"Yuhui Du, Zheng Wang, Ju Niu, Yulong Wang, Godfrey D Pearlson, Vince D Calhoun","doi":"10.1109/TMI.2025.3585880","DOIUrl":"10.1109/TMI.2025.3585880","url":null,"abstract":"<p><p>The subjective nature of diagnosing mental disorders complicates achieving accurate diagnoses. The complex relationship among disorders further exacerbates this issue, particularly in clinical practice where conditions like bipolar disorder (BP) and schizophrenia (SZ) can present similar clinical symptoms and cognitive impairments. To address these challenges, this paper proposes a mutualistic multi-network noisy label learning (MMNNLL) method, which aims to enhance diagnostic accuracy by leveraging neuroimaging data under the presence of potential clinical diagnosis bias or errors. MMNNLL effectively utilizes multiple deep neural networks (DNNs) for learning from data with noisy labels by maximizing the consistency among DNNs in identifying and utilizing samples with clean and noisy labels. Experimental results on public CIFAR-10 and PathMNIST datasets demonstrate the effectiveness of our method in classifying independent test data across various types and levels of label noise. Additionally, our MMNNLL method significantly outperforms state-of-the-art noisy label learning methods. When applied to brain functional connectivity data from BP and SZ patients, our method identifies two biotypes that show more pronounced group differences, and improved classification accuracy compared to the original clinical categories, using both traditional machine learning and advanced deep learning techniques. In summary, our method effectively addresses the possible inaccuracy in nosology of mental disorders and achieves transdiagnostic classification through robust noisy label learning via multi-network collaboration and competition.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":"5014-5026"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12812316/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144565572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Joint Shape Reconstruction and Registration via a Shared Hybrid Diffeomorphic Flow. 基于共享混合差胚流的关节形状重建与配准。
Pub Date : 2025-12-01 DOI: 10.1109/TMI.2025.3585560
Hengxiang Shi, Ping Wang, Shouhui Zhang, Xiuyang Zhao, Bo Yang, Caiming Zhang

Deep implicit functions (DIFs) effectively represent shapes by using a neural network to map 3D spatial coordinates to scalar values that encode the shape's geometry, but it is difficult to establish correspondences between shapes directly, limiting their use in medical image registration. The recently presented deformation field-based methods achieve implicit templates learning via template field learning with DIFs and deformation field learning, establishing shape correspondence through deformation fields. Although these approaches enable joint learning of shape representation and shape correspondence, the decoupled optimization for template field and deformation field, caused by the absence of deformation annotations lead to a relatively accurate template field but an underoptimized deformation field. In this paper, we propose a novel implicit template learning framework via a shared hybrid diffeomorphic flow (SHDF), which enables shared optimization for deformation and template, contributing to better deformations and shape representation. Specifically, we formulate the signed distance function (SDF, a type of DIFs) as a one-dimensional (1D) integral, unifying dimensions to match the form used in solving ordinary differential equation (ODE) for deformation field learning. Then, SDF in 1D integral form is integrated seamlessly into the deformation field learning. Using a recurrent learning strategy, we frame shape representations and deformations as solving different initial value problems of the same ODE. We also introduce a global smoothness regularization to handle local optima due to limited outside-of-shape data. Experiments on medical datasets show that SHDF outperforms state-of-the-art methods in shape representation and registration.

深度隐式函数(Deep implicit functions, dif)利用神经网络将三维空间坐标映射到编码形状几何的标量值,有效地表示形状,但难以直接建立形状之间的对应关系,限制了其在医学图像配准中的应用。最近提出的基于变形场的方法通过模板场学习和变形场学习实现隐式模板学习,通过变形场建立形状对应关系。虽然这些方法可以实现形状表示和形状对应的联合学习,但由于缺乏变形注释,导致模板场和变形场的解耦优化导致模板场相对准确,但变形场优化不足。在本文中,我们提出了一种新的隐式模板学习框架,该框架通过共享混合微分同构流(SHDF)实现变形和模板的共享优化,有助于更好的变形和形状表示。具体来说,我们将有符号距离函数(SDF, dif的一种)表述为一维(1D)积分,统一维度以匹配用于求解变形场学习的常微分方程(ODE)的形式。然后,将一维积分形式的SDF无缝集成到变形场学习中。使用循环学习策略,我们将形状表示和变形框架为解决相同ODE的不同初值问题。我们还引入了全局平滑正则化来处理由于有限的形状外数据而导致的局部最优。在医学数据集上的实验表明,SHDF在形状表示和配准方面优于最先进的方法。
{"title":"Joint Shape Reconstruction and Registration via a Shared Hybrid Diffeomorphic Flow.","authors":"Hengxiang Shi, Ping Wang, Shouhui Zhang, Xiuyang Zhao, Bo Yang, Caiming Zhang","doi":"10.1109/TMI.2025.3585560","DOIUrl":"10.1109/TMI.2025.3585560","url":null,"abstract":"<p><p>Deep implicit functions (DIFs) effectively represent shapes by using a neural network to map 3D spatial coordinates to scalar values that encode the shape's geometry, but it is difficult to establish correspondences between shapes directly, limiting their use in medical image registration. The recently presented deformation field-based methods achieve implicit templates learning via template field learning with DIFs and deformation field learning, establishing shape correspondence through deformation fields. Although these approaches enable joint learning of shape representation and shape correspondence, the decoupled optimization for template field and deformation field, caused by the absence of deformation annotations lead to a relatively accurate template field but an underoptimized deformation field. In this paper, we propose a novel implicit template learning framework via a shared hybrid diffeomorphic flow (SHDF), which enables shared optimization for deformation and template, contributing to better deformations and shape representation. Specifically, we formulate the signed distance function (SDF, a type of DIFs) as a one-dimensional (1D) integral, unifying dimensions to match the form used in solving ordinary differential equation (ODE) for deformation field learning. Then, SDF in 1D integral form is integrated seamlessly into the deformation field learning. Using a recurrent learning strategy, we frame shape representations and deformations as solving different initial value problems of the same ODE. We also introduce a global smoothness regularization to handle local optima due to limited outside-of-shape data. Experiments on medical datasets show that SHDF outperforms state-of-the-art methods in shape representation and registration.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":"4998-5013"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144562424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Guest Editorial Special Issue on Advancements in Foundation Models for Medical Imaging 特邀评论:医学影像基础模型的进展
Pub Date : 2025-10-27 DOI: 10.1109/TMI.2025.3613074
Tianming Liu;Dinggang Shen;Jong Chul Ye;Marleen de Bruijne;Wei Liu
Pretrained on massive datasets, Foundation Models (FMs) are revolutionizing medical imaging by offering scalable and generalizable solutions to longstanding challenges. This Special Issue on Advancements in Foundation Models for Medical Imaging presents FM-related works that explore the potential of FMs to address data scarcity, domain shifts, and multimodal integration across a wide range of medical imaging tasks, including segmentation, diagnosis, reconstruction, and prognosis. The included papers also examine critical concerns such as interpretability, efficiency, benchmarking, and ethics in the adoption of FMs for medical imaging. Collectively, the articles in this Special Issue mark a significant step toward establishing FMs as a cornerstone of next-generation medical imaging AI.
基础模型(FMs)在海量数据集上进行预训练,通过提供可扩展和通用的解决方案来解决长期存在的挑战,正在彻底改变医学成像。本期《医学成像基础模型进展》特刊介绍了与医学成像相关的工作,探讨了医学成像在解决数据短缺、领域转移和跨广泛医学成像任务(包括分割、诊断、重建和预后)的多模式集成方面的潜力。纳入的论文还研究了一些关键问题,如可解释性、效率、基准和医学成像中采用FMs的伦理。总的来说,本期特刊中的文章标志着将FMs作为下一代医学成像人工智能的基石迈出了重要的一步。
{"title":"Guest Editorial Special Issue on Advancements in Foundation Models for Medical Imaging","authors":"Tianming Liu;Dinggang Shen;Jong Chul Ye;Marleen de Bruijne;Wei Liu","doi":"10.1109/TMI.2025.3613074","DOIUrl":"https://doi.org/10.1109/TMI.2025.3613074","url":null,"abstract":"Pretrained on massive datasets, Foundation Models (FMs) are revolutionizing medical imaging by offering scalable and generalizable solutions to longstanding challenges. This Special Issue on Advancements in Foundation Models for Medical Imaging presents FM-related works that explore the potential of FMs to address data scarcity, domain shifts, and multimodal integration across a wide range of medical imaging tasks, including segmentation, diagnosis, reconstruction, and prognosis. The included papers also examine critical concerns such as interpretability, efficiency, benchmarking, and ethics in the adoption of FMs for medical imaging. Collectively, the articles in this Special Issue mark a significant step toward establishing FMs as a cornerstone of next-generation medical imaging AI.","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"44 10","pages":"3894-3897"},"PeriodicalIF":0.0,"publicationDate":"2025-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11218696","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145371487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging Diffusion Model and Image Foundation Model for Improved Correspondence Matching in Coronary Angiography. 利用弥散模型和图像基础模型改进冠状动脉造影的对应匹配。
Pub Date : 2025-10-20 DOI: 10.1109/TMI.2025.3623507
Lin Zhao, Xin Yu, Yikang Liu, Xiao Chen, Eric Z Chen, Terrence Chen, Shanhui Sun

Accurate correspondence matching in coronary angiography images is crucial for reconstructing 3D coronary artery structures, which is essential for precise diagnosis and treatment planning of coronary artery disease (CAD). Traditional matching methods for natural images often fail to generalize to X-ray images due to inherent differences such as lack of texture, lower contrast, and overlapping structures, compounded by insufficient training data. To address these challenges, we propose a novel pipeline that generates realistic paired coronary angiography images using a diffusion model conditioned on 2D projections of 3D reconstructed meshes from Coronary Computed Tomography Angiography (CCTA), providing high-quality synthetic data for training. Additionally, we employ large-scale image foundation models to guide feature aggregation, enhancing correspondence matching accuracy by focusing on semantically relevant regions and keypoints. Our approach demonstrates superior matching performance on synthetic datasets and effectively generalizes to real-world datasets, offering a practical solution for this task. Furthermore, our work investigates the efficacy of different foundation models in correspondence matching, providing novel insights into leveraging advanced image foundation models for medical imaging applications.

冠状动脉造影图像的精确对应匹配是重建冠状动脉三维结构的关键,对冠状动脉疾病(CAD)的精确诊断和治疗计划至关重要。传统的自然图像匹配方法由于缺乏纹理、对比度较低、结构重叠等固有的差异,再加上训练数据不足,往往不能推广到x射线图像。为了解决这些挑战,我们提出了一种新的管道,该管道使用冠状动脉ct血管造影(CCTA)三维重建网格的二维投影为条件的扩散模型,生成逼真的成对冠状动脉造影图像,为训练提供高质量的合成数据。此外,我们采用大规模的图像基础模型来引导特征聚合,通过关注语义相关的区域和关键点来提高对应匹配的准确性。我们的方法在合成数据集上展示了卓越的匹配性能,并有效地推广到现实世界的数据集,为这项任务提供了一个实用的解决方案。此外,我们的工作研究了不同基础模型在对应匹配中的功效,为利用先进的图像基础模型进行医学成像应用提供了新的见解。
{"title":"Leveraging Diffusion Model and Image Foundation Model for Improved Correspondence Matching in Coronary Angiography.","authors":"Lin Zhao, Xin Yu, Yikang Liu, Xiao Chen, Eric Z Chen, Terrence Chen, Shanhui Sun","doi":"10.1109/TMI.2025.3623507","DOIUrl":"https://doi.org/10.1109/TMI.2025.3623507","url":null,"abstract":"<p><p>Accurate correspondence matching in coronary angiography images is crucial for reconstructing 3D coronary artery structures, which is essential for precise diagnosis and treatment planning of coronary artery disease (CAD). Traditional matching methods for natural images often fail to generalize to X-ray images due to inherent differences such as lack of texture, lower contrast, and overlapping structures, compounded by insufficient training data. To address these challenges, we propose a novel pipeline that generates realistic paired coronary angiography images using a diffusion model conditioned on 2D projections of 3D reconstructed meshes from Coronary Computed Tomography Angiography (CCTA), providing high-quality synthetic data for training. Additionally, we employ large-scale image foundation models to guide feature aggregation, enhancing correspondence matching accuracy by focusing on semantically relevant regions and keypoints. Our approach demonstrates superior matching performance on synthetic datasets and effectively generalizes to real-world datasets, offering a practical solution for this task. Furthermore, our work investigates the efficacy of different foundation models in correspondence matching, providing novel insights into leveraging advanced image foundation models for medical imaging applications.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145338362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FairFedMed: Benchmarking Group Fairness in Federated Medical Imaging with FairLoRA. FairFedMed:利用FairLoRA对联邦医学成像中的群体公平性进行基准测试。
Pub Date : 2025-10-16 DOI: 10.1109/TMI.2025.3622522
Minghan Li, Congcong Wen, Yu Tian, Min Shi, Yan Luo, Hao Huang, Yi Fang, Mengyu Wang

Fairness remains a critical concern in healthcare, where unequal access to services and treatment outcomes can adversely affect patient health. While Federated Learning (FL) presents a collaborative and privacy-preserving approach to model training, ensuring fairness is challenging due to heterogeneous data across institutions, and current research primarily addresses non-medical applications. To fill this gap, we establish the first experimental benchmark for fairness in medical FL, evaluating six representative FL methods across diverse demographic attributes and imaging modalities. We introduce FairFedMed, the first medical FL dataset specifically designed to study group fairness (i.e., consistent performance across demographic groups). It comprises two parts: FairFedMed-Oph, featuring 2D fundus and 3D OCT ophthalmology samples with six demographic attributes; and FairFedMed-Chest, which simulates real cross-institutional FL using subsets of CheXpert and MIMIC-CXR. Together, they support both simulated and real-world FL across diverse medical modalities and demographic groups. Existing FL models often underperform on medical images and overlook fairness across demographic groups. To address this, we propose FairLoRA, a fairness-aware FL framework based on SVD-based low-rank approximation. It customizes singular value matrices per demographic group while sharing singular vectors, ensuring both fairness and efficiency. Experimental results on the FairFedMed dataset demonstrate that FairLoRA not only achieves state-of-the-art performance in medical image classification but also significantly improves fairness across diverse populations. Our code and dataset can be accessible via GitHub link: https://github.com/Harvard-AI-and-Robotics-Lab/FairFedMed.

公平仍然是医疗保健领域的一个关键问题,在医疗保健领域,获得服务和治疗结果的机会不平等可能对患者健康产生不利影响。虽然联邦学习(FL)提出了一种协作和隐私保护的模型训练方法,但由于跨机构的异构数据,确保公平性具有挑战性,目前的研究主要针对非医疗应用。为了填补这一空白,我们建立了医疗FL公平性的第一个实验基准,评估了跨越不同人口统计属性和成像方式的六种代表性FL方法。我们介绍了FairFedMed,这是第一个专门用于研究群体公平性(即跨人口统计群体的一致表现)的医疗FL数据集。它包括两个部分:FairFedMed-Oph,包含二维眼底和三维OCT眼科样本,具有六个人口统计学属性;FairFedMed-Chest,使用CheXpert和MIMIC-CXR的子集模拟真实的跨机构FL。总之,他们支持模拟和现实世界的FL跨越不同的医疗模式和人口群体。现有的FL模型通常在医学图像上表现不佳,并且忽略了人口群体之间的公平性。为了解决这个问题,我们提出了FairLoRA,这是一个基于基于svd的低秩近似的公平感知FL框架。它在共享奇异向量的同时,为每个人口群体定制奇异值矩阵,保证了公平性和效率。在FairFedMed数据集上的实验结果表明,FairLoRA不仅在医学图像分类方面达到了最先进的性能,而且显著提高了不同人群的公平性。我们的代码和数据集可以通过GitHub链接访问:https://github.com/Harvard-AI-and-Robotics-Lab/FairFedMed。
{"title":"FairFedMed: Benchmarking Group Fairness in Federated Medical Imaging with FairLoRA.","authors":"Minghan Li, Congcong Wen, Yu Tian, Min Shi, Yan Luo, Hao Huang, Yi Fang, Mengyu Wang","doi":"10.1109/TMI.2025.3622522","DOIUrl":"10.1109/TMI.2025.3622522","url":null,"abstract":"<p><p>Fairness remains a critical concern in healthcare, where unequal access to services and treatment outcomes can adversely affect patient health. While Federated Learning (FL) presents a collaborative and privacy-preserving approach to model training, ensuring fairness is challenging due to heterogeneous data across institutions, and current research primarily addresses non-medical applications. To fill this gap, we establish the first experimental benchmark for fairness in medical FL, evaluating six representative FL methods across diverse demographic attributes and imaging modalities. We introduce FairFedMed, the first medical FL dataset specifically designed to study group fairness (i.e., consistent performance across demographic groups). It comprises two parts: FairFedMed-Oph, featuring 2D fundus and 3D OCT ophthalmology samples with six demographic attributes; and FairFedMed-Chest, which simulates real cross-institutional FL using subsets of CheXpert and MIMIC-CXR. Together, they support both simulated and real-world FL across diverse medical modalities and demographic groups. Existing FL models often underperform on medical images and overlook fairness across demographic groups. To address this, we propose FairLoRA, a fairness-aware FL framework based on SVD-based low-rank approximation. It customizes singular value matrices per demographic group while sharing singular vectors, ensuring both fairness and efficiency. Experimental results on the FairFedMed dataset demonstrate that FairLoRA not only achieves state-of-the-art performance in medical image classification but also significantly improves fairness across diverse populations. Our code and dataset can be accessible via GitHub link: https://github.com/Harvard-AI-and-Robotics-Lab/FairFedMed.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145310409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MoE-Morph: Lightweight Pyramid Model With Heterogeneous Mixture of Experts for Deformable Medical Image Registration MoE-Morph:用于形变医学图像配准的非均匀混合专家轻量级金字塔模型。
Pub Date : 2025-10-14 DOI: 10.1109/TMI.2025.3620406
Hao Lin;Yonghong Song;You Su;Yunfei Ma
Deformable image registration aims to achieve nonlinear alignment of image spaces by estimating dense displacement fields. It is widely used in clinical tasks such as surgical planning, assisted diagnosis, and surgical navigation. While efficient, deep learning registration methods often struggle with large, complex displacements. Pyramid-based approaches address this with a coarse-to-fine strategy, but their single-feature processing can lead to error accumulation. In this paper, we introduce a dense Mixture of Experts (MoE) pyramid registration model, using routing schemes and multiple heterogeneous experts to increase the width and flexibility of feature processing within a single layer. The collaboration among heterogeneous experts enables the model to retain more precise details and maintain greater feature freedom when dealing with complex displacements. We use only deformation fields as the information transmission paradigm between different levels, with deformation field interactions between layers, which encourages the model to focus on the feature location matching process and perform registration in the correct direction. We do not utilize any complex mechanisms such as attention or ViT, keeping the model at its simplest form. The powerful deformable capability allows the model to perform volume registration directly and accurately without the need for affine registration. Experimental results show that the model achieves outstanding performance across four public datasets, including brain registration, lung registration, and abdominal multi-modal registration. The code will be published at https://github.com/Darlinglinlinlin/MOE_Morph
变形图像配准是通过估计密集位移场来实现图像空间的非线性对齐。它被广泛应用于临床任务,如手术计划、辅助诊断和手术导航。虽然高效,但深度学习的注册方法往往难以处理大而复杂的位移。基于金字塔的方法通过一种从粗到精的策略来解决这个问题,但是它们的单一特征处理可能导致错误积累。在本文中,我们引入了密集混合专家(MoE)金字塔配准模型,使用路由方案和多个异构专家来增加单层内特征处理的宽度和灵活性。异构专家之间的协作使模型能够在处理复杂位移时保留更精确的细节并保持更大的特征自由度。我们只使用变形场作为不同层次之间的信息传递范式,层与层之间存在变形场的相互作用,这促使模型专注于特征位置匹配过程,并朝着正确的方向进行配准。我们不使用任何复杂的机制,如注意力或ViT,使模型保持最简单的形式。强大的可变形能力使模型可以直接准确地进行体配准,而无需进行仿射配准。实验结果表明,该模型在脑配准、肺配准和腹部多模态配准四个公共数据集上都取得了优异的性能。代码将在https://github.com/Darlinglinlinlin/MOE_Morph上发布。
{"title":"MoE-Morph: Lightweight Pyramid Model With Heterogeneous Mixture of Experts for Deformable Medical Image Registration","authors":"Hao Lin;Yonghong Song;You Su;Yunfei Ma","doi":"10.1109/TMI.2025.3620406","DOIUrl":"10.1109/TMI.2025.3620406","url":null,"abstract":"Deformable image registration aims to achieve nonlinear alignment of image spaces by estimating dense displacement fields. It is widely used in clinical tasks such as surgical planning, assisted diagnosis, and surgical navigation. While efficient, deep learning registration methods often struggle with large, complex displacements. Pyramid-based approaches address this with a coarse-to-fine strategy, but their single-feature processing can lead to error accumulation. In this paper, we introduce a dense Mixture of Experts (MoE) pyramid registration model, using routing schemes and multiple heterogeneous experts to increase the width and flexibility of feature processing within a single layer. The collaboration among heterogeneous experts enables the model to retain more precise details and maintain greater feature freedom when dealing with complex displacements. We use only deformation fields as the information transmission paradigm between different levels, with deformation field interactions between layers, which encourages the model to focus on the feature location matching process and perform registration in the correct direction. We do not utilize any complex mechanisms such as attention or ViT, keeping the model at its simplest form. The powerful deformable capability allows the model to perform volume registration directly and accurately without the need for affine registration. Experimental results show that the model achieves outstanding performance across four public datasets, including brain registration, lung registration, and abdominal multi-modal registration. The code will be published at <uri>https://github.com/Darlinglinlinlin/MOE_Morph</uri>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"45 3","pages":"1251-1264"},"PeriodicalIF":0.0,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145288378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncertainty-Guided Prototype Reliability Enhancement Network for Few-Shot Medical Image Segmentation 基于不确定性的少镜头医学图像分割原型可靠性增强网络。
Pub Date : 2025-10-14 DOI: 10.1109/TMI.2025.3621452
Junfei Hu;Tao Zhou;Kaiwen Huang;Yi Zhou;Haofeng Zhang;Boqiang Fan;Huazhu Fu
Few-Shot Learning (FSL) has garnered increasing attention for data-scarce scenarios, particularly in medical segmentation tasks where only a few labeled data points are available. Existing few-shot segmentation methods typically learn prototypes from support images and employ nearest-neighbor searching to segment query images. Despite notable progress, effectively learning prototypes for each class remains a challenging task to achieve promising results. In this paper, we propose an Uncertainty-guided Prototype Reliability Enhancement Network (UPRE-Net) for few-shot medical image segmentation. Specifically, we present a dual-support branch to maximize the extraction of information from support images through augmentation techniques. To enhance the reliability of prototypes, we propose an Uncertainty-guided Prototype Generation (UPG) module. Within the UPG module, we first extract both global and local prototypes for each class and then apply uncertainty measures to select the most informative prototypes. Additionally, to effectively combine the prediction results from the dual-support branch, we present a Reliable Dynamic Fusion (RDF) module. This module dynamically integrates the two prediction results to generate a more reliable output. Furthermore, we present an Uncertainty-induced Weighted Loss (UWL) to ensure that the model pays more attention to these regions with high uncertainty. Experiments on four benchmark medical image datasets demonstrate that our proposed model significantly outperforms state-of-the-art methods. The code will be released at https://github.com/taozh2017/UPRENet
在数据稀缺的情况下,特别是在只有少数标记数据点可用的医疗分割任务中,few - shot Learning (FSL)获得了越来越多的关注。现有的小镜头分割方法通常是从支持图像中学习原型,并采用最近邻搜索对查询图像进行分割。尽管取得了显著的进展,但有效地学习每个类的原型仍然是一项具有挑战性的任务,以实现有希望的结果。本文提出了一种基于不确定性的原型可靠性增强网络(UPRE-Net),用于医学图像分割。具体来说,我们提出了一个双支持分支,通过增强技术最大限度地从支持图像中提取信息。为了提高原型的可靠性,我们提出了一种不确定性引导的原型生成(UPG)模块。在UPG模块中,我们首先为每个类提取全局和局部原型,然后应用不确定性度量来选择信息最多的原型。此外,为了有效地结合双支持分支的预测结果,我们提出了可靠动态融合(RDF)模块。该模块动态整合两种预测结果,生成更可靠的输出。此外,我们提出了不确定性加权损失(UWL),以确保模型更加关注这些具有高不确定性的区域。在四个基准医学图像数据集上的实验表明,我们提出的模型明显优于最先进的方法。代码将在https://github.com/taozh2017/UPRENet上发布。
{"title":"Uncertainty-Guided Prototype Reliability Enhancement Network for Few-Shot Medical Image Segmentation","authors":"Junfei Hu;Tao Zhou;Kaiwen Huang;Yi Zhou;Haofeng Zhang;Boqiang Fan;Huazhu Fu","doi":"10.1109/TMI.2025.3621452","DOIUrl":"10.1109/TMI.2025.3621452","url":null,"abstract":"Few-Shot Learning (FSL) has garnered increasing attention for data-scarce scenarios, particularly in medical segmentation tasks where only a few labeled data points are available. Existing few-shot segmentation methods typically learn prototypes from support images and employ nearest-neighbor searching to segment query images. Despite notable progress, effectively learning prototypes for each class remains a challenging task to achieve promising results. In this paper, we propose an Uncertainty-guided Prototype Reliability Enhancement Network (UPRE-Net) for few-shot medical image segmentation. Specifically, we present a dual-support branch to maximize the extraction of information from support images through augmentation techniques. To enhance the reliability of prototypes, we propose an Uncertainty-guided Prototype Generation (UPG) module. Within the UPG module, we first extract both global and local prototypes for each class and then apply uncertainty measures to select the most informative prototypes. Additionally, to effectively combine the prediction results from the dual-support branch, we present a Reliable Dynamic Fusion (RDF) module. This module dynamically integrates the two prediction results to generate a more reliable output. Furthermore, we present an Uncertainty-induced Weighted Loss (UWL) to ensure that the model pays more attention to these regions with high uncertainty. Experiments on four benchmark medical image datasets demonstrate that our proposed model significantly outperforms state-of-the-art methods. The code will be released at <uri>https://github.com/taozh2017/UPRENet</uri>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"45 3","pages":"1279-1290"},"PeriodicalIF":0.0,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145288544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PET Head Motion Estimation Using Supervised Deep Learning With Attention 基于监督深度学习的PET头部运动估计
Pub Date : 2025-10-13 DOI: 10.1109/TMI.2025.3620714
Zhuotong Cai;Tianyi Zeng;Jiazhen Zhang;Eléonore V. Lieffrig;Kathryn Fontaine;Chenyu You;Enette Mae Revilla;James S. Duncan;Jingmin Xin;Yihuan Lu;John A. Onofrey
Head movement poses a significant challenge in brain positron emission tomography (PET) imaging, resulting in image artifacts and tracer uptake quantification inaccuracies. Effective head motion estimation and correction are crucial for precise quantitative image analysis and accurate diagnosis of neurological disorders. Hardware-based motion tracking (HMT) has limited applicability in real-world clinical practice. To overcome this limitation, we propose a deep-learning head motion correction approach with cross-attention (DL-HMC++) to predict rigid head motion from one-second 3D PET raw data. DL-HMC++ is trained in a supervised manner by leveraging existing dynamic PET scans with gold-standard motion measurements from external HMT. We evaluate DL-HMC++ on two PET scanners (HRRT and mCT) and four radiotracers (18F-FDG, 18F-FPEB, 11C-UCB-J, and 11C-LSN3172176) to demonstrate the effectiveness and generalization of the approach in large cohort PET studies. Quantitative and qualitative results demonstrate that DL-HMC++ consistently outperforms state-of-the-art data-driven motion estimation methods, producing motion-free images with clear delineation of brain structures and reduced motion artifacts that are indistinguishable from gold-standard HMT. Brain region of interest standard uptake value analysis exhibits average difference ratios between DL-HMC++ and gold-standard HMT to be $1.2pm 0.5$ % for HRRT and $0.5pm 0.2$ % for mCT. DL-HMC++ demonstrates the potential for data-driven PET head motion correction to remove the burden of HMT, making motion correction accessible to clinical populations beyond research settings. The code is available at https://github.com/maxxxxxxcai/DL-HMC-TMI
头部运动对脑正电子发射断层扫描(PET)成像提出了重大挑战,导致图像伪影和示踪剂摄取量化不准确。有效的头部运动估计和校正对于精确定量图像分析和准确诊断神经系统疾病至关重要。基于硬件的运动跟踪(HMT)在现实世界的临床实践中适用性有限。为了克服这一限制,我们提出了一种基于交叉注意的深度学习头部运动校正方法(dl - hmc++),从一秒3D PET原始数据中预测刚性头部运动。dl - hmc++以监督的方式训练,利用现有的动态PET扫描与外部HMT的金标准运动测量。我们在两种PET扫描仪(HRRT和mCT)和四种放射性示踪剂(18F-FDG, 18F-FPEB, 11C-UCB-J和11C-LSN3172176)上评估dl - hmc++,以证明该方法在大型队列PET研究中的有效性和泛化性。定量和定性结果表明,dl - hmc++始终优于最先进的数据驱动运动估计方法,产生无运动图像,清晰描绘大脑结构,减少运动伪影,与金标准HMT无法区分。脑兴趣区标准摄取值分析显示dl - hmc++与金标准HMT之间的平均差异比为:HRRT为$1.2pm 0.5$ %, mCT为$0.5pm 0.2$ %。dl - hmc++展示了数据驱动的PET头部运动校正的潜力,以消除HMT的负担,使运动校正可以在研究环境之外的临床人群中使用。代码可在https://github.com/maxxxxxxcai/DL-HMC-TMI上获得
{"title":"PET Head Motion Estimation Using Supervised Deep Learning With Attention","authors":"Zhuotong Cai;Tianyi Zeng;Jiazhen Zhang;Eléonore V. Lieffrig;Kathryn Fontaine;Chenyu You;Enette Mae Revilla;James S. Duncan;Jingmin Xin;Yihuan Lu;John A. Onofrey","doi":"10.1109/TMI.2025.3620714","DOIUrl":"10.1109/TMI.2025.3620714","url":null,"abstract":"Head movement poses a significant challenge in brain positron emission tomography (PET) imaging, resulting in image artifacts and tracer uptake quantification inaccuracies. Effective head motion estimation and correction are crucial for precise quantitative image analysis and accurate diagnosis of neurological disorders. Hardware-based motion tracking (HMT) has limited applicability in real-world clinical practice. To overcome this limitation, we propose a deep-learning head motion correction approach with cross-attention (DL-HMC++) to predict rigid head motion from one-second 3D PET raw data. DL-HMC++ is trained in a supervised manner by leveraging existing dynamic PET scans with gold-standard motion measurements from external HMT. We evaluate DL-HMC++ on two PET scanners (HRRT and mCT) and four radiotracers (<sup>18</sup>F-FDG, <sup>18</sup>F-FPEB, <sup>11</sup>C-UCB-J, and <sup>11</sup>C-LSN3172176) to demonstrate the effectiveness and generalization of the approach in large cohort PET studies. Quantitative and qualitative results demonstrate that DL-HMC++ consistently outperforms state-of-the-art data-driven motion estimation methods, producing motion-free images with clear delineation of brain structures and reduced motion artifacts that are indistinguishable from gold-standard HMT. Brain region of interest standard uptake value analysis exhibits average difference ratios between DL-HMC++ and gold-standard HMT to be <inline-formula> <tex-math>$1.2pm 0.5$ </tex-math></inline-formula>% for HRRT and <inline-formula> <tex-math>$0.5pm 0.2$ </tex-math></inline-formula>% for mCT. DL-HMC++ demonstrates the potential for data-driven PET head motion correction to remove the burden of HMT, making motion correction accessible to clinical populations beyond research settings. The code is available at <uri>https://github.com/maxxxxxxcai/DL-HMC-TMI</uri>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"45 3","pages":"1265-1278"},"PeriodicalIF":0.0,"publicationDate":"2025-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145282738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised High-Order Implicit Neural Representation With Line Attention for Metal Artifact Reduction 基于线注意的无监督高阶隐式神经表示用于金属伪影还原
Pub Date : 2025-10-10 DOI: 10.1109/TMI.2025.3620222
Hongyu Chen;Shaoguang Huang;Wei He;Guangyi Yang;Hongyan Zhang
The presence of metallic implants introduces bright and dark streaks that appear in computed tomography (CT) images, degrading image quality and interfering with medical diagnosis. To reduce these artifacts, deep learning approaches have been applied for metal-corrupted restoration, which usually requires a large amount of simulated degraded-clean pairs for training. To achieve metal artifact reduction (MAR) without reference images, implicit neural representation (INR) has emerged and shown capabilities for image restoration in an unsupervised manner. However, existing INR methods for MAR usually treat the spatial coordinates independently and ignore their correlation, resulting in detail loss and artifacts remaining. In this paper, we propose an INR-based unsupervised MAR framework and design a High-order Line Attention Network to capture local contextual and geometric representations from X-rays, which maps the spatial coordinates into discrete linear attenuation coefficients of imaged objects for artifact-free CT image reconstruction. The second-order feature interaction can effectively improve the spectral bias problems and fit low and high-frequency details of real signals well. The proposed line-attention module with linear complexity can establish global relationships among spatial point tokens from sampled rays. To provide more local contextual information, a multiple local adjacent ray sampling strategy is adopted to compose several sub-fan beams with more context as a training batch. With the help of these components, the unsupervised MAR framework can approximate the implicit continuous function to estimate measurements and generate artifact-free CT images. Simulated and real experiments indicated that the proposed approach achieved superior MAR performance compared with other state-of-the-art methods.
金属植入物的存在会在计算机断层扫描(CT)图像中出现明亮和黑暗的条纹,降低图像质量并干扰医学诊断。为了减少这些伪影,深度学习方法已被应用于金属损坏修复,这通常需要大量模拟的退化清洁对进行训练。为了在没有参考图像的情况下实现金属伪影还原(MAR),隐式神经表征(INR)已经出现并显示出以无监督方式恢复图像的能力。然而,现有的用于MAR的INR方法通常对空间坐标进行独立处理,忽略它们之间的相关性,导致细节丢失和伪影的存在。在本文中,我们提出了一个基于inr的无监督MAR框架,并设计了一个高阶线注意网络来捕获x射线的局部上下文和几何表示,该网络将空间坐标映射到成像物体的离散线性衰减系数中,用于无伪影CT图像重建。二阶特征交互可以有效改善频谱偏置问题,并能很好地拟合真实信号的低频和高频细节。所提出的线性复杂度线性关注模块可以在采样射线的空间点标记之间建立全局关系。为了提供更多的局部上下文信息,采用多局部相邻射线采样策略,组成多个具有更多上下文的子扇波束作为训练批。在这些分量的帮助下,无监督MAR框架可以近似隐式连续函数来估计测量值并生成无伪影的CT图像。仿真和实际实验表明,该方法具有较好的MAR性能。
{"title":"Unsupervised High-Order Implicit Neural Representation With Line Attention for Metal Artifact Reduction","authors":"Hongyu Chen;Shaoguang Huang;Wei He;Guangyi Yang;Hongyan Zhang","doi":"10.1109/TMI.2025.3620222","DOIUrl":"10.1109/TMI.2025.3620222","url":null,"abstract":"The presence of metallic implants introduces bright and dark streaks that appear in computed tomography (CT) images, degrading image quality and interfering with medical diagnosis. To reduce these artifacts, deep learning approaches have been applied for metal-corrupted restoration, which usually requires a large amount of simulated degraded-clean pairs for training. To achieve metal artifact reduction (MAR) without reference images, implicit neural representation (INR) has emerged and shown capabilities for image restoration in an unsupervised manner. However, existing INR methods for MAR usually treat the spatial coordinates independently and ignore their correlation, resulting in detail loss and artifacts remaining. In this paper, we propose an INR-based unsupervised MAR framework and design a High-order Line Attention Network to capture local contextual and geometric representations from X-rays, which maps the spatial coordinates into discrete linear attenuation coefficients of imaged objects for artifact-free CT image reconstruction. The second-order feature interaction can effectively improve the spectral bias problems and fit low and high-frequency details of real signals well. The proposed line-attention module with linear complexity can establish global relationships among spatial point tokens from sampled rays. To provide more local contextual information, a multiple local adjacent ray sampling strategy is adopted to compose several sub-fan beams with more context as a training batch. With the help of these components, the unsupervised MAR framework can approximate the implicit continuous function to estimate measurements and generate artifact-free CT images. Simulated and real experiments indicated that the proposed approach achieved superior MAR performance compared with other state-of-the-art methods.","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"45 3","pages":"1237-1250"},"PeriodicalIF":0.0,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145260753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on medical imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1