首页 > 最新文献

IEEE transactions on medical imaging最新文献

英文 中文
PAINT: Prior-Aided Alternate Iterative NeTwork for Ultra-Low-Dose CT Imaging Using Diffusion Model-Restored Sinogram. PAINT:基于扩散模型恢复Sinogram超低剂量CT成像的先验辅助交替迭代网络。
Pub Date : 2026-02-01 DOI: 10.1109/TMI.2025.3599508
Kaile Chen, Weikang Zhang, Ziheng Deng, Yufu Zhou, Jun Zhao

Obtaining multiple CT scans from the same patient is required in many clinical scenarios, such as lung nodule screening and image-guided radiation therapy. Repeated scans would expose patients to higher radiation dose and increase the risk of cancer. In this study, we aim to achieve ultra-low-dose imaging for subsequent scans by collecting extremely undersampled sinogram via regional few-view scanning, and preserve image quality utilizing the preceding fullsampled scan as prior. To fully exploit prior information, we propose a two-stage framework consisting of diffusion model-based sinogram restoration and deep learning-based unrolled iterative reconstruction. Specifically, the undersampled sinogram is first restored by a conditional diffusion model with sinogram-domain prior guidance. Then, we formulate the undersampled data reconstruction problem as an optimization problem combining fidelity terms for both undersampled and restored data, along with a regularization term based on image-domain prior. Next, we propose Prior-aided Alternate Iterative NeTwork (PAINT) to solve the optimization problem. PAINT alternately updates the undersampled or restored data fidelity term, and unrolls the iterations to integrate neural network-based prior regularization. In the case of 112 mm field of view in simulated data experiments, our proposed framework achieved superior performance in terms of CT value accuracy and image details preservation. Clinical data experiments also demonstrated that our proposed framework outperformed the comparison methods in artifact reduction and structure recovery.

在许多临床情况下,需要对同一患者进行多次CT扫描,例如肺结节筛查和图像引导放射治疗。重复扫描会使病人暴露在更高的辐射剂量下,增加患癌症的风险。在本研究中,我们的目标是通过区域少视点扫描收集极度欠采样的正弦图,为后续扫描实现超低剂量成像,并利用之前的全采样扫描保持图像质量。为了充分利用先验信息,我们提出了一个两阶段框架,包括基于扩散模型的正弦图恢复和基于深度学习的展开迭代重建。具体地说,欠采样的正弦图首先通过带有正弦图域先验引导的条件扩散模型恢复。然后,我们将欠采样数据重建问题表述为将欠采样和恢复数据的保真度项以及基于图像域先验的正则化项结合在一起的优化问题。接下来,我们提出了先验辅助交替迭代网络(PAINT)来解决优化问题。PAINT交替更新欠采样或恢复的数据保真度项,并展开迭代以集成基于神经网络的先验正则化。在模拟数据实验中,在112 mm视场的情况下,我们提出的框架在CT值精度和图像细节保存方面取得了优异的性能。临床数据实验也表明,我们提出的框架在伪影减少和结构恢复方面优于比较方法。
{"title":"PAINT: Prior-Aided Alternate Iterative NeTwork for Ultra-Low-Dose CT Imaging Using Diffusion Model-Restored Sinogram.","authors":"Kaile Chen, Weikang Zhang, Ziheng Deng, Yufu Zhou, Jun Zhao","doi":"10.1109/TMI.2025.3599508","DOIUrl":"10.1109/TMI.2025.3599508","url":null,"abstract":"<p><p>Obtaining multiple CT scans from the same patient is required in many clinical scenarios, such as lung nodule screening and image-guided radiation therapy. Repeated scans would expose patients to higher radiation dose and increase the risk of cancer. In this study, we aim to achieve ultra-low-dose imaging for subsequent scans by collecting extremely undersampled sinogram via regional few-view scanning, and preserve image quality utilizing the preceding fullsampled scan as prior. To fully exploit prior information, we propose a two-stage framework consisting of diffusion model-based sinogram restoration and deep learning-based unrolled iterative reconstruction. Specifically, the undersampled sinogram is first restored by a conditional diffusion model with sinogram-domain prior guidance. Then, we formulate the undersampled data reconstruction problem as an optimization problem combining fidelity terms for both undersampled and restored data, along with a regularization term based on image-domain prior. Next, we propose Prior-aided Alternate Iterative NeTwork (PAINT) to solve the optimization problem. PAINT alternately updates the undersampled or restored data fidelity term, and unrolls the iterations to integrate neural network-based prior regularization. In the case of 112 mm field of view in simulated data experiments, our proposed framework achieved superior performance in terms of CT value accuracy and image details preservation. Clinical data experiments also demonstrated that our proposed framework outperformed the comparison methods in artifact reduction and structure recovery.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":"434-447"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144877631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ultrasound Autofocusing: Common Midpoint Phase Error Optimization via Differentiable Beamforming. 超声自动聚焦:基于可微波束形成的常见中点相位误差优化。
Pub Date : 2026-02-01 DOI: 10.1109/TMI.2025.3607875
Walter Simson, Louise Zhuang, Benjamin N Frey, Sergio J Sanabria, Jeremy J Dahl, Dongwoon Hyun

In ultrasound imaging, propagation of an acoustic wavefront through heterogeneous media causes phase aberrations that degrade the coherence of the reflected wavefront, leading to reduced image resolution and contrast. Adaptive imaging techniques attempt to correct this phase aberration and restore coherence, leading to improved focusing of the image. We propose an autofocusing paradigm for aberration correction in ultrasound imaging by fitting an acoustic velocity field to pressure measurements, via optimization of the common midpoint phase error (CMPE), using a straight-ray wave propagation model for beamforming in diffusely scattering media. We show that CMPE induced by heterogeneous acoustic velocity is a robust measure of phase aberration that can be used for acoustic autofocusing. CMPE is optimized iteratively using a differentiable beamforming approach to simultaneously improve the image focus while estimating the acoustic velocity field of the interrogated medium. The approach relies solely on wavefield measurements using a straight-ray integral solution of the two-way time-of-flight without explicit numerical time-stepping models of wave propagation. We demonstrate method performance through in silico simulations, in vitro phantom measurements, and in vivo mammalian models, showing practical applications in distributed aberration quantification, correction, and velocity estimation for medical ultrasound autofocusing.

在超声成像中,声波前通过异质介质的传播会引起相位像差,从而降低反射波前的相干性,导致图像分辨率和对比度降低。自适应成像技术试图纠正这种相位像差并恢复相干性,从而改善图像的聚焦。我们提出了一种自动聚焦模式,通过优化共同中点相位误差(CMPE),将声速场拟合到压力测量值中,从而校正超声成像中的像差,并使用漫射散射介质中波束形成的直线波传播模型。研究表明,由非均匀声速引起的CMPE是一种可靠的相位像差测量方法,可用于声学自动聚焦。利用可微波束形成方法对CMPE进行迭代优化,在估计被测介质声速场的同时提高了图像聚焦。该方法完全依赖于使用双向飞行时间的直线积分解的波场测量,而没有明确的波传播数值时间步进模型。我们通过硅模拟、体外幻影测量和体内哺乳动物模型展示了该方法的性能,展示了在医学超声自动聚焦的分布式像差量化、校正和速度估计方面的实际应用。
{"title":"Ultrasound Autofocusing: Common Midpoint Phase Error Optimization via Differentiable Beamforming.","authors":"Walter Simson, Louise Zhuang, Benjamin N Frey, Sergio J Sanabria, Jeremy J Dahl, Dongwoon Hyun","doi":"10.1109/TMI.2025.3607875","DOIUrl":"10.1109/TMI.2025.3607875","url":null,"abstract":"<p><p>In ultrasound imaging, propagation of an acoustic wavefront through heterogeneous media causes phase aberrations that degrade the coherence of the reflected wavefront, leading to reduced image resolution and contrast. Adaptive imaging techniques attempt to correct this phase aberration and restore coherence, leading to improved focusing of the image. We propose an autofocusing paradigm for aberration correction in ultrasound imaging by fitting an acoustic velocity field to pressure measurements, via optimization of the common midpoint phase error (CMPE), using a straight-ray wave propagation model for beamforming in diffusely scattering media. We show that CMPE induced by heterogeneous acoustic velocity is a robust measure of phase aberration that can be used for acoustic autofocusing. CMPE is optimized iteratively using a differentiable beamforming approach to simultaneously improve the image focus while estimating the acoustic velocity field of the interrogated medium. The approach relies solely on wavefield measurements using a straight-ray integral solution of the two-way time-of-flight without explicit numerical time-stepping models of wave propagation. We demonstrate method performance through in silico simulations, in vitro phantom measurements, and in vivo mammalian models, showing practical applications in distributed aberration quantification, correction, and velocity estimation for medical ultrasound autofocusing.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":"681-692"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145031617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ring artifact reduction in photon counting CT using redundant sampling and autocalibration. 利用冗余采样和自动校准减少光子计数CT中的环形伪影。
Pub Date : 2026-01-26 DOI: 10.1109/TMI.2026.3658004
Scott S Hsieh, James Day, Xinchen Deng, Magdalena Bazalova-Carter

Ring artifacts in CT are caused by uncalibrated variations in detector pixels and are especially prevalent with emerging photon counting detectors (PCDs). Control of ring artifacts is conventionally accomplished by improving either hardware manufacturing or software correction algorithms. An alternative solution is detector autocalibration, in which two redundant samples of each line integral are acquired and used to dynamically calibrate the PCD. Autocalibration was first proposed by Hounsfield in 1977 and was demonstrated on the EMI Topaz prototype scanner in 1980, but details surrounding this implementation are sparse. We investigate a form of autocalibration that requires just two redundant acquisitions, which could be acquired using flying focal spot on a clinical scanner but is demonstrated here with a detector shift. We formulated autocalibration as an optimization problem to determine the relative gain factor of each pixel and tested it on scans of a chicken thigh specimen, resolution phantom, and a cylindrical phantom. Ring artifacts were significantly reduced. Some residual artifacts remained but could not be discriminated from the intrinsic temporal instability of our PCD modules. Autocalibration could facilitate the adoption of widespread photon counting CT by reducing ring artifacts, thermal management requirements, or stability requirements that are present today. Demonstration of autocalibration on a rotating gantry with flying focal spot remains future work.

CT中的环形伪影是由探测器像素的未校准变化引起的,在新兴的光子计数探测器(PCDs)中尤其普遍。环形工件的控制通常是通过改进硬件制造或软件校正算法来完成的。另一种解决方案是探测器自动校准,其中每个线积分获得两个冗余样本并用于动态校准PCD。自动校准最初是由Hounsfield于1977年提出的,并于1980年在EMI Topaz原型扫描仪上进行了演示,但围绕该实现的细节很少。我们研究了一种自动校准的形式,它只需要两个冗余的采集,可以使用临床扫描仪上的飞行焦点来获得,但这里用检测器移位来演示。我们将自动校准制定为一个优化问题,以确定每个像素的相对增益因子,并在鸡大腿标本、分辨率幻影和圆柱形幻影的扫描上进行了测试。环形伪影显著减少。一些残留的工件仍然存在,但不能从我们的PCD模块固有的时间不稳定性中区分出来。通过减少目前存在的环形伪影、热管理要求或稳定性要求,自动校准可以促进光子计数CT的广泛采用。在具有飞行焦斑的旋转龙门上进行自动校准的演示仍有待进一步的工作。
{"title":"Ring artifact reduction in photon counting CT using redundant sampling and autocalibration.","authors":"Scott S Hsieh, James Day, Xinchen Deng, Magdalena Bazalova-Carter","doi":"10.1109/TMI.2026.3658004","DOIUrl":"https://doi.org/10.1109/TMI.2026.3658004","url":null,"abstract":"<p><p>Ring artifacts in CT are caused by uncalibrated variations in detector pixels and are especially prevalent with emerging photon counting detectors (PCDs). Control of ring artifacts is conventionally accomplished by improving either hardware manufacturing or software correction algorithms. An alternative solution is detector autocalibration, in which two redundant samples of each line integral are acquired and used to dynamically calibrate the PCD. Autocalibration was first proposed by Hounsfield in 1977 and was demonstrated on the EMI Topaz prototype scanner in 1980, but details surrounding this implementation are sparse. We investigate a form of autocalibration that requires just two redundant acquisitions, which could be acquired using flying focal spot on a clinical scanner but is demonstrated here with a detector shift. We formulated autocalibration as an optimization problem to determine the relative gain factor of each pixel and tested it on scans of a chicken thigh specimen, resolution phantom, and a cylindrical phantom. Ring artifacts were significantly reduced. Some residual artifacts remained but could not be discriminated from the intrinsic temporal instability of our PCD modules. Autocalibration could facilitate the adoption of widespread photon counting CT by reducing ring artifacts, thermal management requirements, or stability requirements that are present today. Demonstration of autocalibration on a rotating gantry with flying focal spot remains future work.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146055735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neuron Segment Connectivity Prediction with Multimodal Features for Connectomics. 基于多模态特征的连接组学神经元连接预测。
Pub Date : 2026-01-26 DOI: 10.1109/TMI.2026.3658169
Qihua Chen, Xuejin Chen, Chenxuan Wang, Zhiwei Xiong, Feng Wu

Reconstructing neurons from large electron microscopy (EM) datasets for connectomic analysis presents a significant challenge, particularly in segmenting neurons of complex morphologies. Previous deep learning-based neuron segmentation methods often rely on pixel-level image context and produce extensive oversegmented fragments. Detecting these split errors and merging the split neuron segments are non-trivial for various neurons in a large-scale EM data volume. In this work, we exploit multimodal features in the full workflow of automatic neuron proofreading. We propose a novel connection point detection network that utilizes both global 3D morphological features and high-resolution local image context to extract candidate segment pairs from massive adjacent segments. To effectively fuse the 3D morphological feature and the dense image features from very different scales, we design a proposal-based image feature sampling to improve the efficiency of multimodal cross-attentions. Integrating the connection point detection network with our connectivity prediction network which also utilizes multimodal features, we make a fully automatic neuron segment merging pipeline, closely imitating human proofreading. Comprehensive experimental results verify the effectiveness of the proposed modules and demonstrate the robustness of the entire pipeline in large-scale neuron reconstruction. The code and data are available at https://github.com/Levishery/ Neuron-Segment-Connection-Prediction.

从大型电子显微镜(EM)数据集重建神经元用于连接组分析提出了重大挑战,特别是在分割复杂形态的神经元时。以往基于深度学习的神经元分割方法往往依赖于像素级图像上下文,产生大量的过分割碎片。在大规模的EM数据量中,检测这些分裂错误并合并分裂神经元片段是非常重要的。在这项工作中,我们在自动神经元校对的整个工作流程中利用了多模态特征。我们提出了一种新的连接点检测网络,该网络利用全局三维形态特征和高分辨率的局部图像上下文从大量相邻片段中提取候选片段对。为了有效地融合三维形态特征和不同尺度的密集图像特征,设计了一种基于提议的图像特征采样方法,提高了多模态交叉关注的效率。将连接点检测网络与利用多模态特征的连通性预测网络相结合,构建了一个模仿人工校对的全自动神经元段合并流水线。综合实验结果验证了所提模块的有效性,并证明了整个管道在大规模神经元重建中的鲁棒性。代码和数据可在https://github.com/Levishery/神经元-段-连接-预测。
{"title":"Neuron Segment Connectivity Prediction with Multimodal Features for Connectomics.","authors":"Qihua Chen, Xuejin Chen, Chenxuan Wang, Zhiwei Xiong, Feng Wu","doi":"10.1109/TMI.2026.3658169","DOIUrl":"https://doi.org/10.1109/TMI.2026.3658169","url":null,"abstract":"<p><p>Reconstructing neurons from large electron microscopy (EM) datasets for connectomic analysis presents a significant challenge, particularly in segmenting neurons of complex morphologies. Previous deep learning-based neuron segmentation methods often rely on pixel-level image context and produce extensive oversegmented fragments. Detecting these split errors and merging the split neuron segments are non-trivial for various neurons in a large-scale EM data volume. In this work, we exploit multimodal features in the full workflow of automatic neuron proofreading. We propose a novel connection point detection network that utilizes both global 3D morphological features and high-resolution local image context to extract candidate segment pairs from massive adjacent segments. To effectively fuse the 3D morphological feature and the dense image features from very different scales, we design a proposal-based image feature sampling to improve the efficiency of multimodal cross-attentions. Integrating the connection point detection network with our connectivity prediction network which also utilizes multimodal features, we make a fully automatic neuron segment merging pipeline, closely imitating human proofreading. Comprehensive experimental results verify the effectiveness of the proposed modules and demonstrate the robustness of the entire pipeline in large-scale neuron reconstruction. The code and data are available at https://github.com/Levishery/ Neuron-Segment-Connection-Prediction.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146055699","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Highly Undersampled MRI Reconstruction via a Single Posterior Sampling of Diffusion Models. 通过扩散模型的单一后验采样进行高度欠采样MRI重建。
Pub Date : 2026-01-16 DOI: 10.1109/TMI.2026.3654585
Jin Liu, Qing Lin, Zhuang Xiong, Shanshan Shan, Chunyi Liu, Min Li, Feng Liu, G Bruce Pike, Hongfu Sun, Yang Gao

Incoherent k-space undersampling and deep learning-based reconstruction methods have shown great success in accelerating MRI. However, the performance of most previous methods will degrade dramatically under high acceleration factors, e.g., 8× or higher. Recently, denoising diffusion models (DM) have demonstrated promising results in solving this issue; however, one major drawback of the DM methods is the long inference time due to a dramatic number of iterative reverse posterior sampling steps. In this work, a Single Step Diffusion Model-based reconstruction framework, namely SSDM-MRI, is proposed for restoring MRI images from highly undersampled k-space. The proposed method achieves one-step reconstruction by first training a conditional DM and then iteratively distilling this model four times using an iterative selective distillation algorithm, which works synergistically with a shortcut reverse sampling strategy for model inference. Comprehensive experiments were carried out on both publicly available fastMRI brain and knee images, as well as an in-house multi-echo GRE (QSM) subject. Overall, the results showed that SSDM-MRI outperformed other methods in terms of numerical metrics (e.g., PSNR and SSIM), error maps, image fine details, and latent susceptibility information hidden in MRI phase images. In addition, the reconstruction time for a 320×320 brain slice of SSDM-MRI is only 0.45 second, which is only comparable to that of a simple U-net, making it a highly effective solution for MRI reconstruction tasks.

非相干k空间欠采样和基于深度学习的重建方法在加速MRI方面取得了巨大成功。然而,大多数以前的方法在高加速因素下的性能会急剧下降,例如8倍或更高。近年来,扩散去噪模型(DM)在解决这一问题上取得了可喜的成果;然而,DM方法的一个主要缺点是由于大量的迭代反向后验采样步骤导致推理时间长。在这项工作中,提出了一种基于单步扩散模型的重建框架,即SSDM-MRI,用于从高度欠采样的k空间中恢复MRI图像。该方法首先训练一个条件DM,然后使用迭代选择蒸馏算法对该模型进行四次迭代蒸馏,从而实现一步重建,并与模型推理的快捷反向采样策略协同工作。综合实验是在公开的快速mri脑和膝关节图像以及内部多回声GRE (QSM)受试者上进行的。总体而言,结果表明SSDM-MRI在数值指标(如PSNR和SSIM)、误差图、图像精细细节和MRI相位图像中隐藏的潜在敏感性信息等方面优于其他方法。此外,SSDM-MRI的320×320脑切片重建时间仅为0.45秒,仅与简单的U-net相当,是MRI重建任务的高效解决方案。
{"title":"Highly Undersampled MRI Reconstruction via a Single Posterior Sampling of Diffusion Models.","authors":"Jin Liu, Qing Lin, Zhuang Xiong, Shanshan Shan, Chunyi Liu, Min Li, Feng Liu, G Bruce Pike, Hongfu Sun, Yang Gao","doi":"10.1109/TMI.2026.3654585","DOIUrl":"https://doi.org/10.1109/TMI.2026.3654585","url":null,"abstract":"<p><p>Incoherent k-space undersampling and deep learning-based reconstruction methods have shown great success in accelerating MRI. However, the performance of most previous methods will degrade dramatically under high acceleration factors, e.g., 8× or higher. Recently, denoising diffusion models (DM) have demonstrated promising results in solving this issue; however, one major drawback of the DM methods is the long inference time due to a dramatic number of iterative reverse posterior sampling steps. In this work, a Single Step Diffusion Model-based reconstruction framework, namely SSDM-MRI, is proposed for restoring MRI images from highly undersampled k-space. The proposed method achieves one-step reconstruction by first training a conditional DM and then iteratively distilling this model four times using an iterative selective distillation algorithm, which works synergistically with a shortcut reverse sampling strategy for model inference. Comprehensive experiments were carried out on both publicly available fastMRI brain and knee images, as well as an in-house multi-echo GRE (QSM) subject. Overall, the results showed that SSDM-MRI outperformed other methods in terms of numerical metrics (e.g., PSNR and SSIM), error maps, image fine details, and latent susceptibility information hidden in MRI phase images. In addition, the reconstruction time for a 320×320 brain slice of SSDM-MRI is only 0.45 second, which is only comparable to that of a simple U-net, making it a highly effective solution for MRI reconstruction tasks.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145991983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Regression is all you need for medical image translation. 回归是所有你需要的医学图像翻译。
Pub Date : 2026-01-01 DOI: 10.1109/TMI.2025.3650412
Sebastian Rassmann, David Kugler, Christian Ewert, Martin Reuter

While Generative Adversarial Nets (GANs) and Diffusion Models (DMs) have achieved impressive results in natural image synthesis, their core strengths - creativity and realism - can be detrimental in medical applications, where accuracy and fidelity are paramount. These models instead risk introducing hallucinations and replication of unwanted acquisition noise. Here, we propose YODA (You Only Denoise once - or Average), a 2.5D diffusion-based framework for medical image translation (MIT). Consistent with DM theory, we find that conventional diffusion sampling stochastically replicates noise. To mitigate this, we draw and average multiple samples, akin to physical signal averaging. As this effectively approximates the DM's expected value, we term this Expectation-Approximation (ExpA) sampling. We additionally propose regression sampling YODA, which retains the initial DM prediction and omits iterative refinement to produce noise-free images in a single step. Across five diverse multi-modal datasets - including multi-contrast brain MRI and pelvic MRI-CT - we demonstrate that regression sampling is not only substantially more efficient but also matches or exceeds image quality of full diffusion sampling even with ExpA. Our results reveal that iterative refinement solely enhances perceptual realism without benefiting information translation, which we confirm in relevant downstream tasks. YODA outperforms eight state-of-the-art DMs and GANs and challenges the presumed superiority of DMs and GANs over computationally cheap regression models for high-quality MIT. Furthermore, we show that YODA-translated images are interchangeable with, or even superior to, physical acquisitions for several medical applications.

虽然生成对抗网络(gan)和扩散模型(dm)在自然图像合成方面取得了令人印象深刻的成果,但它们的核心优势——创造力和真实感——在准确性和保真度至关重要的医疗应用中可能是有害的。相反,这些模型冒着引入幻觉和复制不必要的获取噪音的风险。在这里,我们提出了YODA(你只去噪一次或平均),一个基于2.5D扩散的医学图像翻译框架(MIT)。与DM理论一致,我们发现传统的扩散采样随机地复制了噪声。为了减轻这种情况,我们绘制并平均多个样本,类似于物理信号平均。由于这有效地近似DM的期望值,我们称之为期望-近似(ExpA)抽样。我们还提出了回归采样YODA,它保留了初始DM预测并省略了迭代改进,从而在单步中产生无噪声图像。通过五种不同的多模态数据集(包括多对比脑MRI和骨盆MRI- ct),我们证明回归采样不仅更有效,而且即使使用ExpA也可以匹配或超过完全扩散采样的图像质量。我们的研究结果表明,迭代细化只增强了感知真实感,而对信息翻译没有好处,我们在相关的下游任务中证实了这一点。YODA超越了8个最先进的dm和gan,并挑战了dm和gan相对于高质量MIT计算成本低廉的回归模型的假定优势。此外,我们证明了yoda翻译的图像与几种医学应用的物理采集可以互换,甚至优于物理采集。
{"title":"Regression is all you need for medical image translation.","authors":"Sebastian Rassmann, David Kugler, Christian Ewert, Martin Reuter","doi":"10.1109/TMI.2025.3650412","DOIUrl":"https://doi.org/10.1109/TMI.2025.3650412","url":null,"abstract":"<p><p>While Generative Adversarial Nets (GANs) and Diffusion Models (DMs) have achieved impressive results in natural image synthesis, their core strengths - creativity and realism - can be detrimental in medical applications, where accuracy and fidelity are paramount. These models instead risk introducing hallucinations and replication of unwanted acquisition noise. Here, we propose YODA (You Only Denoise once - or Average), a 2.5D diffusion-based framework for medical image translation (MIT). Consistent with DM theory, we find that conventional diffusion sampling stochastically replicates noise. To mitigate this, we draw and average multiple samples, akin to physical signal averaging. As this effectively approximates the DM's expected value, we term this Expectation-Approximation (ExpA) sampling. We additionally propose regression sampling YODA, which retains the initial DM prediction and omits iterative refinement to produce noise-free images in a single step. Across five diverse multi-modal datasets - including multi-contrast brain MRI and pelvic MRI-CT - we demonstrate that regression sampling is not only substantially more efficient but also matches or exceeds image quality of full diffusion sampling even with ExpA. Our results reveal that iterative refinement solely enhances perceptual realism without benefiting information translation, which we confirm in relevant downstream tasks. YODA outperforms eight state-of-the-art DMs and GANs and challenges the presumed superiority of DMs and GANs over computationally cheap regression models for high-quality MIT. Furthermore, we show that YODA-translated images are interchangeable with, or even superior to, physical acquisitions for several medical applications.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145890774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmark of Segmentation Techniques for Pelvic Fracture in CT and X-Ray: Summary of the PENGWIN 2024 Challenge. 骨盆骨折CT和x线分割技术的标杆:PENGWIN 2024挑战综述
Pub Date : 2026-01-01 DOI: 10.1109/TMI.2025.3650126
Yudi Sang, Yanzhen Liu, Sutuke Yibulayimu, Yunning Wang, Benjamin D Killeen, Mingxu Liu, Ping-Cheng Ku, Ole Johannsen, Karol Gotkowski, Maximilian Zenk, Klaus Maier-Hein, Fabian Isensee, Peiyan Yue, Yi Wang, Haidong Yu, Zhaohong Pan, Yutong He, Xiaokun Liang, Daiqi Liu, Fuxin Fan, Artur Jurgas, Andrzej Skalski, Yuxi Ma, Jing Yang, Szymon Plotka, Rafal Litka, Gang Zhu, Yingchun Song, Mathias Unberath, Mehran Armand, Dan Ruan, S Kevin Zhou, Qiyong Cao, Chunpeng Zhao, Xinbao Wu, Yu Wang

The segmentation of pelvic fracture fragments in CT and X-ray images is crucial for trauma diagnosis, surgical planning, and intraoperative guidance. However, accurately and efficiently delineating the bone fragments remains a significant challenge due to complex anatomy and imaging limitations. The PENGWIN challenge, organized as a MICCAI 2024 satellite event, aimed to advance automated fracture segmentation by benchmarking state-of-the-art algorithms on these complex tasks. A diverse dataset of 150 CT scans was collected from multiple clinical centers, and a large set of simulated X-ray images was generated using the DeepDRR method. Final submissions from 16 teams worldwide were evaluated under a rigorous multi-metric testing scheme. The top-performing CT algorithm achieved an average fragment-wise intersection over union (IoU) of 0.930, demonstrating satisfactory accuracy. However, in the X-ray task, the best algorithm achieved an IoU of 0.774, which is promising but not yet sufficient for intra-operative decision-making, reflecting the inherent challenges of fragment overlap in projection imaging. Beyond the quantitative evaluation, the challenge revealed methodological diversity in algorithm design. Variations in instance representation, such as primary-secondary classification versus boundary-core separation, led to differing segmentation strategies. Despite promising results, the challenge also exposed inherent uncertainties in fragment definition, particularly in cases of incomplete fractures. These findings suggest that interactive segmentation approaches, integrating human decision-making with task-relevant information, may be essential for improving model reliability and clinical applicability.

骨盆骨折碎片的CT和x线图像分割对创伤诊断、手术计划和术中指导至关重要。然而,由于复杂的解剖结构和成像限制,准确有效地描绘骨碎片仍然是一个重大挑战。PENGWIN挑战赛作为MICCAI 2024卫星活动组织,旨在通过对这些复杂任务的最先进算法进行基准测试来推进自动化裂缝分割。从多个临床中心收集了150个CT扫描的不同数据集,并使用DeepDRR方法生成了大量模拟x射线图像。来自全球16个团队的最终作品在严格的多指标测试计划下进行了评估。表现最好的CT算法获得了0.930的平均分段交叉优于联合(IoU),显示出令人满意的精度。然而,在x射线任务中,最佳算法的IoU为0.774,这对于术中决策是有希望的,但还不够,这反映了投影成像中碎片重叠的固有挑战。除了定量评估之外,挑战还揭示了算法设计方法的多样性。实例表示的变化,例如主要-次要分类与边界-核心分离,导致了不同的分割策略。尽管取得了令人鼓舞的结果,但挑战也暴露了碎片定义的固有不确定性,特别是在不完全骨折的情况下。这些发现表明,将人类决策与任务相关信息相结合的交互式分割方法可能对提高模型可靠性和临床适用性至关重要。
{"title":"Benchmark of Segmentation Techniques for Pelvic Fracture in CT and X-Ray: Summary of the PENGWIN 2024 Challenge.","authors":"Yudi Sang, Yanzhen Liu, Sutuke Yibulayimu, Yunning Wang, Benjamin D Killeen, Mingxu Liu, Ping-Cheng Ku, Ole Johannsen, Karol Gotkowski, Maximilian Zenk, Klaus Maier-Hein, Fabian Isensee, Peiyan Yue, Yi Wang, Haidong Yu, Zhaohong Pan, Yutong He, Xiaokun Liang, Daiqi Liu, Fuxin Fan, Artur Jurgas, Andrzej Skalski, Yuxi Ma, Jing Yang, Szymon Plotka, Rafal Litka, Gang Zhu, Yingchun Song, Mathias Unberath, Mehran Armand, Dan Ruan, S Kevin Zhou, Qiyong Cao, Chunpeng Zhao, Xinbao Wu, Yu Wang","doi":"10.1109/TMI.2025.3650126","DOIUrl":"https://doi.org/10.1109/TMI.2025.3650126","url":null,"abstract":"<p><p>The segmentation of pelvic fracture fragments in CT and X-ray images is crucial for trauma diagnosis, surgical planning, and intraoperative guidance. However, accurately and efficiently delineating the bone fragments remains a significant challenge due to complex anatomy and imaging limitations. The PENGWIN challenge, organized as a MICCAI 2024 satellite event, aimed to advance automated fracture segmentation by benchmarking state-of-the-art algorithms on these complex tasks. A diverse dataset of 150 CT scans was collected from multiple clinical centers, and a large set of simulated X-ray images was generated using the DeepDRR method. Final submissions from 16 teams worldwide were evaluated under a rigorous multi-metric testing scheme. The top-performing CT algorithm achieved an average fragment-wise intersection over union (IoU) of 0.930, demonstrating satisfactory accuracy. However, in the X-ray task, the best algorithm achieved an IoU of 0.774, which is promising but not yet sufficient for intra-operative decision-making, reflecting the inherent challenges of fragment overlap in projection imaging. Beyond the quantitative evaluation, the challenge revealed methodological diversity in algorithm design. Variations in instance representation, such as primary-secondary classification versus boundary-core separation, led to differing segmentation strategies. Despite promising results, the challenge also exposed inherent uncertainties in fragment definition, particularly in cases of incomplete fractures. These findings suggest that interactive segmentation approaches, integrating human decision-making with task-relevant information, may be essential for improving model reliability and clinical applicability.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145890767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NeeCo: Image Synthesis of Novel Instrument States Based on Dynamic and Deformable 3D Gaussian Reconstruction. 基于动态和可变形三维高斯重构的新型仪器状态图像合成。
Pub Date : 2025-12-30 DOI: 10.1109/TMI.2025.3648299
Tianle Zeng, Junlei Hu, Gerardo Loza Galindo, Sharib Ali, Duygu Sarikaya, Pietro Valdastri, Dominic Jones

Computer vision-based technologies significantly enhance surgical automation by advancing tool tracking, detection, and localization. However, Current data-driven approaches are data-voracious, requiring large, high-quality labeled image datasets. Our Work introduces a novel dynamic Gaussian Splatting technique to address the data scarcity in surgical image datasets. We propose a dynamic Gaussian model to represent dynamic surgical scenes, enabling the rendering of surgical instruments from unseen viewpoints and deformations with real tissue backgrounds. We utilize a dynamic training adjustment strategy to address challenges posed by poorly calibrated camera poses from real-world scenarios. Additionally, automatically generate annotations for our synthetic data. For evaluation, we constructed a new dataset featuring seven scenes with 14,000 frames of tool and camera motion and tool jaw articulation, with a background of an exvivo porcine model. Using this dataset, we synthetically replicate the scene deformation from the ground truth data, allowing direct comparisons of synthetic image quality. Experimental results illustrate that our method generates photo-realistic labeled image datasets with the highest PSNR (29.87). We further evaluate the performance of medical-specific neural networks trained on real and synthetic images using an unseen real-world image dataset. Our results show that the performance of models trained on synthetic images generated by the proposed method outperforms those trained with state-of-the-art standard data augmentation by 10%, leading to an overall improvement in model performances by nearly 15%.

基于计算机视觉的技术通过推进工具跟踪、检测和定位,显著提高了手术自动化程度。然而,当前的数据驱动方法是海量数据,需要大量高质量的标记图像数据集。我们的工作介绍了一种新的动态高斯飞溅技术来解决手术图像数据集的数据稀缺性。我们提出了一个动态高斯模型来表示动态手术场景,使手术器械从看不见的角度和真实组织背景的变形呈现。我们利用动态训练调整策略来解决现实世界场景中校准不良的相机姿势所带来的挑战。此外,为合成数据自动生成注释。为了进行评估,我们构建了一个新的数据集,该数据集包含7个场景,包含14,000帧工具和相机运动以及工具颚关节,背景为体外猪模型。使用该数据集,我们从地面真实数据中合成复制场景变形,从而可以直接比较合成图像质量。实验结果表明,我们的方法生成了具有最高PSNR(29.87)的逼真的标记图像数据集。我们使用未见过的真实世界图像数据集进一步评估在真实和合成图像上训练的医疗特定神经网络的性能。我们的研究结果表明,使用该方法生成的合成图像训练的模型的性能比使用最先进的标准数据增强训练的模型的性能高出10%,导致模型性能的总体提高了近15%。
{"title":"NeeCo: Image Synthesis of Novel Instrument States Based on Dynamic and Deformable 3D Gaussian Reconstruction.","authors":"Tianle Zeng, Junlei Hu, Gerardo Loza Galindo, Sharib Ali, Duygu Sarikaya, Pietro Valdastri, Dominic Jones","doi":"10.1109/TMI.2025.3648299","DOIUrl":"https://doi.org/10.1109/TMI.2025.3648299","url":null,"abstract":"<p><p>Computer vision-based technologies significantly enhance surgical automation by advancing tool tracking, detection, and localization. However, Current data-driven approaches are data-voracious, requiring large, high-quality labeled image datasets. Our Work introduces a novel dynamic Gaussian Splatting technique to address the data scarcity in surgical image datasets. We propose a dynamic Gaussian model to represent dynamic surgical scenes, enabling the rendering of surgical instruments from unseen viewpoints and deformations with real tissue backgrounds. We utilize a dynamic training adjustment strategy to address challenges posed by poorly calibrated camera poses from real-world scenarios. Additionally, automatically generate annotations for our synthetic data. For evaluation, we constructed a new dataset featuring seven scenes with 14,000 frames of tool and camera motion and tool jaw articulation, with a background of an exvivo porcine model. Using this dataset, we synthetically replicate the scene deformation from the ground truth data, allowing direct comparisons of synthetic image quality. Experimental results illustrate that our method generates photo-realistic labeled image datasets with the highest PSNR (29.87). We further evaluate the performance of medical-specific neural networks trained on real and synthetic images using an unseen real-world image dataset. Our results show that the performance of models trained on synthetic images generated by the proposed method outperforms those trained with state-of-the-art standard data augmentation by 10%, leading to an overall improvement in model performances by nearly 15%.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145866783","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anatomy-aware Sketch-guided Latent Diffusion Model for Orbital Tumor Multi-Parametric MRI Missing Modalities Synthesis. 眼眶肿瘤多参数MRI缺失模态合成的解剖感知草图引导潜伏扩散模型。
Pub Date : 2025-12-29 DOI: 10.1109/TMI.2025.3648852
Langtao Zhou, Xiaoxia Qu, Tianyu Fu, Jiaoyang Wu, Hong Song, Jingfan Fan, Danni Ai, Deqiang Xiao, Junfang Xian, Jian Yang

Synthesizing missing modalities in multi-parametric MRI (mpMRI) is vital for accurate tumor diagnosis, yet remains challenging due to incomplete acquisitions and modality heterogeneity. Diffusion models have shown strong generative capability, but conventional approaches typically operate in the image domain with high memory costs and often rely solely on noise-space supervision, which limits anatomical fidelity. Latent diffusion models (LDMs) improve efficiency by performing denoising in latent space, but standard LDMs lack explicit structural priors and struggle to integrate multiple modalities effectively. To address these limitations, we propose the anatomy-aware sketch-guided latent diffusion model (ASLDM), a novel LDM-based framework designed for flexible and structure-preserving MRI synthesis. ASLDM incorporates an anatomy-aware feature fusion module, which encodes tumor region masks and edge-based anatomical sketches via cross-attention to guide the denoising process with explicit structure priors. A modality synergistic reconstruction strategy enables the joint modeling of available and missing modalities, enhancing cross-modal consistency and supporting arbitrary missing scenarios. Additionally, we introduce image-level losses for pixel-space supervision using L1 and SSIM losses, overcoming the limitations of pure noise-based loss training and improving the anatomical accuracy of synthesized outputs. Extensive experiments on a five-modality orbital tumor mpMRI private dataset and a four-modality public BraTS2024 dataset demonstrate that ASLDM outperforms state-of-the-art methods in both synthesis quality and structural consistency, showing strong potential for clinically reliable multi-modal MRI completion. Our code is publicly available at: https://github.com/zltshadow/ASLDM.git.

综合多参数MRI (mpMRI)中的缺失模式对于准确诊断肿瘤至关重要,但由于采集不完整和模式异质性,仍然具有挑战性。扩散模型显示出强大的生成能力,但传统方法通常在图像域中操作,具有较高的存储成本,并且通常仅依赖于噪声空间监督,这限制了解剖保真度。潜在扩散模型(ldm)通过在潜在空间中进行去噪来提高效率,但标准ldm缺乏明确的结构先验,难以有效地集成多个模型。为了解决这些限制,我们提出了解剖学感知草图引导潜在扩散模型(ASLDM),这是一种基于ldm的新型框架,旨在实现灵活且保留结构的MRI合成。ASLDM包含一个解剖感知特征融合模块,该模块通过交叉注意对肿瘤区域掩模和基于边缘的解剖草图进行编码,以明确的结构先验指导去噪过程。模态协同重建策略可以对可用模态和缺失模态进行联合建模,增强跨模态一致性并支持任意缺失场景。此外,我们使用L1和SSIM损失引入图像级损失进行像素空间监督,克服了纯基于噪声的损失训练的局限性,提高了合成输出的解剖精度。在五模态轨道肿瘤mpMRI私有数据集和四模态BraTS2024公共数据集上进行的大量实验表明,ASLDM在合成质量和结构一致性方面都优于最先进的方法,显示出临床可靠的多模态MRI完成的强大潜力。我们的代码可以在https://github.com/zltshadow/ASLDM.git上公开获得。
{"title":"Anatomy-aware Sketch-guided Latent Diffusion Model for Orbital Tumor Multi-Parametric MRI Missing Modalities Synthesis.","authors":"Langtao Zhou, Xiaoxia Qu, Tianyu Fu, Jiaoyang Wu, Hong Song, Jingfan Fan, Danni Ai, Deqiang Xiao, Junfang Xian, Jian Yang","doi":"10.1109/TMI.2025.3648852","DOIUrl":"https://doi.org/10.1109/TMI.2025.3648852","url":null,"abstract":"<p><p>Synthesizing missing modalities in multi-parametric MRI (mpMRI) is vital for accurate tumor diagnosis, yet remains challenging due to incomplete acquisitions and modality heterogeneity. Diffusion models have shown strong generative capability, but conventional approaches typically operate in the image domain with high memory costs and often rely solely on noise-space supervision, which limits anatomical fidelity. Latent diffusion models (LDMs) improve efficiency by performing denoising in latent space, but standard LDMs lack explicit structural priors and struggle to integrate multiple modalities effectively. To address these limitations, we propose the anatomy-aware sketch-guided latent diffusion model (ASLDM), a novel LDM-based framework designed for flexible and structure-preserving MRI synthesis. ASLDM incorporates an anatomy-aware feature fusion module, which encodes tumor region masks and edge-based anatomical sketches via cross-attention to guide the denoising process with explicit structure priors. A modality synergistic reconstruction strategy enables the joint modeling of available and missing modalities, enhancing cross-modal consistency and supporting arbitrary missing scenarios. Additionally, we introduce image-level losses for pixel-space supervision using L1 and SSIM losses, overcoming the limitations of pure noise-based loss training and improving the anatomical accuracy of synthesized outputs. Extensive experiments on a five-modality orbital tumor mpMRI private dataset and a four-modality public BraTS2024 dataset demonstrate that ASLDM outperforms state-of-the-art methods in both synthesis quality and structural consistency, showing strong potential for clinically reliable multi-modal MRI completion. Our code is publicly available at: https://github.com/zltshadow/ASLDM.git.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145859643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Contrastive Graph Modeling for Cross-Domain Few-Shot Medical Image Segmentation. 跨域少镜头医学图像分割的对比图建模。
Pub Date : 2025-12-29 DOI: 10.1109/TMI.2025.3649239
Yuntian Bo, Tao Zhou, Zechao Li, Haofeng Zhang, Ling Shao

Cross-domain few-shot medical image segmentation (CD-FSMIS) offers a promising and data-efficient solution for medical applications where annotations are severely scarce and multimodal analysis is required. However, existing methods typically filter out domain-specific information to improve generalization, which inadvertently limits cross-domain performance and degrades source-domain accuracy. To address this, we present Contrastive Graph Modeling (C-Graph), a framework that leverages the structural consistency of medical images as a reliable domain-transferable prior. We represent image features as graphs, with pixels as nodes and semantic affinities as edges. A Structural Prior Graph (SPG) layer is proposed to capture and transfer target-category node dependencies and enable global structure modeling through explicit node interactions. Building upon SPG layers, we introduce a Subgraph Matching Decoding (SMD) mechanism that exploits semantic relations among nodes to guide prediction. Furthermore, we design a Confusion-minimizing Node Contrast (CNC) loss to mitigate node ambiguity and sub-graph heterogeneity by contrastively enhancing node discriminability in the graph space. Our method significantly outperforms prior CD-FSMIS approaches across multiple cross-domain benchmarks, achieving state-of-the-art performance while simultaneously preserving strong segmentation accuracy on the source domain. Our code is available at https://github.com/primebo1/C-Graph.

跨域少镜头医学图像分割(CD-FSMIS)为注释严重缺乏和需要多模态分析的医疗应用提供了一种有前途的数据高效解决方案。然而,现有的方法通常会过滤掉特定于领域的信息来提高泛化,这无意中限制了跨领域的性能并降低了源域的准确性。为了解决这个问题,我们提出了对比图建模(C-Graph),这是一个利用医学图像的结构一致性作为可靠的领域可转移先验的框架。我们将图像特征表示为图,像素作为节点,语义亲和力作为边。提出了一种结构先验图(Structural Prior Graph, SPG)层,用于捕获和传递目标类别节点依赖关系,并通过显式节点交互实现全局结构建模。在SPG层的基础上,我们引入了一种子图匹配解码(SMD)机制,该机制利用节点之间的语义关系来指导预测。此外,我们设计了一个最小化混淆的节点对比(CNC)损失,通过对比增强图空间中的节点可辨别性来减轻节点模糊性和子图异质性。我们的方法在多个跨域基准测试中显著优于先前的CD-FSMIS方法,在获得最先进性能的同时,在源域上保持了很强的分割精度。我们的代码可在https://github.com/primebo1/C-Graph上获得。
{"title":"Contrastive Graph Modeling for Cross-Domain Few-Shot Medical Image Segmentation.","authors":"Yuntian Bo, Tao Zhou, Zechao Li, Haofeng Zhang, Ling Shao","doi":"10.1109/TMI.2025.3649239","DOIUrl":"https://doi.org/10.1109/TMI.2025.3649239","url":null,"abstract":"<p><p>Cross-domain few-shot medical image segmentation (CD-FSMIS) offers a promising and data-efficient solution for medical applications where annotations are severely scarce and multimodal analysis is required. However, existing methods typically filter out domain-specific information to improve generalization, which inadvertently limits cross-domain performance and degrades source-domain accuracy. To address this, we present Contrastive Graph Modeling (C-Graph), a framework that leverages the structural consistency of medical images as a reliable domain-transferable prior. We represent image features as graphs, with pixels as nodes and semantic affinities as edges. A Structural Prior Graph (SPG) layer is proposed to capture and transfer target-category node dependencies and enable global structure modeling through explicit node interactions. Building upon SPG layers, we introduce a Subgraph Matching Decoding (SMD) mechanism that exploits semantic relations among nodes to guide prediction. Furthermore, we design a Confusion-minimizing Node Contrast (CNC) loss to mitigate node ambiguity and sub-graph heterogeneity by contrastively enhancing node discriminability in the graph space. Our method significantly outperforms prior CD-FSMIS approaches across multiple cross-domain benchmarks, achieving state-of-the-art performance while simultaneously preserving strong segmentation accuracy on the source domain. Our code is available at https://github.com/primebo1/C-Graph.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145859678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on medical imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1