首页 > 最新文献

IEEE transactions on medical imaging最新文献

英文 中文
M2PL-GAN: Multi-View Multi-Level Pathology Semantic Perception Learning for H&E-to-IHC Virtual Staining. M2PL-GAN: H&E-to-IHC虚拟染色的多视图多层次病理语义感知学习。
Pub Date : 2026-02-26 DOI: 10.1109/TMI.2026.3668248
Zequn Liu, Liangkuan Zhu, Yining Xie, Xiaoqing Hu, Ziyu Zhang, Jing Zhao, Haochen Qi, Jiajun Chen, Jiayi Ma

Immunohistochemistry (IHC) staining is crucial for determining tumor subtypes, obtaining protein expression information, and developing personalized treatment plans. But compared with hematoxylin and eosin (H&E) staining, IHC staining is more complex and expensive. With the advancement of deep learning, converting H&E stained images into IHC stained images has gradually emerged as a solution for obtaining IHC staining. However, current virtual staining processes suffer from difficulties in aligning pathological semantic features, posing significant challenges for network training, which poses significant challenges for network training. To solve these issues, we propose a multi-view multi-level pathology semantic perception learning method for H&E-to-IHC virtual staining (M2PL-GAN). Unlike prior approaches, M2PL-GAN introduces a comprehensive semantic learning paradigm from three views: structural contextual relations, feature distribution, and topology-aware fine-grained semantics. These correspond to the Context-aware Correlation Mechanism (CACM), the Local-aware Distribution Alignment Mechanism (LDAM), and the Graph-aware Bidirectional Contrastive Learning Mechanism (GBCLM) respectively. Among them, CACM enhances contextual consistency by establishing semantic correlations between virtual and real IHC images at local scales. LDAM ensures alignment of semantic feature distributions between virtual and real IHC images, mitigating semantic shifts caused by HE-IHC staining. GBCLM leverages graph neural network to capture topology-aware semantic representations and optimizes semantic feature alignment through bidirectional contrastive learning. Extensive experiments on both public and private datasets demonstrate that our method outperforms state-of-the-art approaches in both quantitative metrics and qualitative evaluations. Our code is available in https://github.com/Pikachu-one/M2PL-GAN.

免疫组织化学(IHC)染色对于确定肿瘤亚型、获得蛋白质表达信息和制定个性化治疗计划至关重要。但与苏木精和伊红(H&E)染色相比,IHC染色更为复杂和昂贵。随着深度学习的推进,将H&E染色图像转换为IHC染色图像逐渐成为获得IHC染色的解决方案。然而,目前的虚拟染色过程在病理语义特征对齐方面存在困难,这给网络训练带来了重大挑战,这也给网络训练带来了重大挑战。为了解决这些问题,我们提出了一种H&E-to-IHC虚拟染色(M2PL-GAN)的多视图多层次病理语义感知学习方法。与之前的方法不同,M2PL-GAN从三个方面引入了一个全面的语义学习范式:结构上下文关系、特征分布和拓扑感知的细粒度语义。这些机制分别对应于上下文感知的相关机制(ccm)、局部感知的分布对齐机制(LDAM)和图感知的双向对比学习机制(GBCLM)。其中,ccm通过在局部尺度上建立虚拟和真实IHC图像之间的语义关联来增强上下文一致性。LDAM确保虚拟和真实IHC图像之间的语义特征分布对齐,减轻HE-IHC染色引起的语义偏移。GBCLM利用图神经网络捕获拓扑感知的语义表示,并通过双向对比学习优化语义特征对齐。在公共和私人数据集上进行的大量实验表明,我们的方法在定量指标和定性评估方面都优于最先进的方法。我们的代码可在https://github.com/Pikachu-one/M2PL-GAN中获得。
{"title":"M2PL-GAN: Multi-View Multi-Level Pathology Semantic Perception Learning for H&E-to-IHC Virtual Staining.","authors":"Zequn Liu, Liangkuan Zhu, Yining Xie, Xiaoqing Hu, Ziyu Zhang, Jing Zhao, Haochen Qi, Jiajun Chen, Jiayi Ma","doi":"10.1109/TMI.2026.3668248","DOIUrl":"https://doi.org/10.1109/TMI.2026.3668248","url":null,"abstract":"<p><p>Immunohistochemistry (IHC) staining is crucial for determining tumor subtypes, obtaining protein expression information, and developing personalized treatment plans. But compared with hematoxylin and eosin (H&E) staining, IHC staining is more complex and expensive. With the advancement of deep learning, converting H&E stained images into IHC stained images has gradually emerged as a solution for obtaining IHC staining. However, current virtual staining processes suffer from difficulties in aligning pathological semantic features, posing significant challenges for network training, which poses significant challenges for network training. To solve these issues, we propose a multi-view multi-level pathology semantic perception learning method for H&E-to-IHC virtual staining (M2PL-GAN). Unlike prior approaches, M2PL-GAN introduces a comprehensive semantic learning paradigm from three views: structural contextual relations, feature distribution, and topology-aware fine-grained semantics. These correspond to the Context-aware Correlation Mechanism (CACM), the Local-aware Distribution Alignment Mechanism (LDAM), and the Graph-aware Bidirectional Contrastive Learning Mechanism (GBCLM) respectively. Among them, CACM enhances contextual consistency by establishing semantic correlations between virtual and real IHC images at local scales. LDAM ensures alignment of semantic feature distributions between virtual and real IHC images, mitigating semantic shifts caused by HE-IHC staining. GBCLM leverages graph neural network to capture topology-aware semantic representations and optimizes semantic feature alignment through bidirectional contrastive learning. Extensive experiments on both public and private datasets demonstrate that our method outperforms state-of-the-art approaches in both quantitative metrics and qualitative evaluations. Our code is available in https://github.com/Pikachu-one/M2PL-GAN.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147313738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Biomechanics-informed Non-rigid Medical Image Registration with Elasticity Theories. 基于弹性理论的生物力学非刚性医学图像配准。
Pub Date : 2026-02-23 DOI: 10.1109/TMI.2026.3665279
Zhe Min, Zachary M C Baum, Shaheer U Saeed, Shixing Ma, Xinzhe Du, Mark Emberton, Dean C Barratt, Zeike A Taylor, Yipeng Hu

Biomechanical modelling of soft tissue provides a method for constraining medical image registration, such that the estimated spatial transformation is considered biophysically plausible. Existing methods either directly optimize the loss function containing the biomechanical-constrained regularization term over deformations, which takes much computational time, or are trained using biomechanically plausible data generated via finite element simulation, which is cumbersome. This work first instantiates the recently-proposed physics-informed neural networks (PINNs) to 3D elastic models that are used to establish the partial differential equations (PDEs) representing physics laws of biomechanical constraints to be satisfied. The registration algorithm that aligns point sets considering PINN-imposed biomechanics (i.e., the forward problem) is then formulated. In addition, the inverse problem and its algorithm of physical parameter (i.e., material property) estimation along with the registration are also formulated and developed. We carefully compare linear and nonlinear elasticity theories' capabilities in solving both tasks of forward registration and inverse physical parameter identification under PINNs respectively. Furthermore, two specific network configurations that leverage one common branch or two individual branches to predict deformation vectors and biomechanical states are also constructed and compared. The proposed PINNs-based registration approaches have been extensively evaluated with three experiments, that is single and multiple patient MRI-US registration using clinical MRI-US pairs, and registration using pairs of undeformed MR images from clinical cases of prostate cancer biopsy and deformed counterparts with finite-element-computed ground-truth deformation. Results demonstrate that the proposed methods achieve state-of-the-art performances compared to biomechanical-model-based and learning-based registration approaches, and the biomechanical constraints of soft tissues have been successfully warranted after registration. The codes are available at https://github.com/ZheMin-1992/Registration_PINNs.

软组织的生物力学建模提供了一种约束医学图像配准的方法,这样估计的空间变换被认为是生物物理上可信的。现有的方法要么直接优化包含生物力学约束正则化项的损失函数,这需要大量的计算时间,要么使用通过有限元模拟生成的生物力学可信数据进行训练,这很麻烦。这项工作首先将最近提出的物理信息神经网络(pinn)实例化为3D弹性模型,该模型用于建立代表生物力学约束的物理定律的偏微分方程(pde)。然后制定了考虑pinn施加的生物力学(即正演问题)的对齐点集的配准算法。此外,还制定和发展了配准过程中物理参数(即材料属性)估计的反问题及其算法。我们仔细比较了线性和非线性弹性理论分别解决pinn下正向配准和逆物理参数识别任务的能力。此外,还构建和比较了利用一个公共分支或两个单独分支来预测变形向量和生物力学状态的两种特定网络配置。提出的基于pnas的配准方法已经通过三个实验进行了广泛的评估,即使用临床MRI-US对进行单个和多个患者的MRI-US配准,以及使用来自前列腺癌活检临床病例的未变形MR图像对和具有有限元计算的地基真值变形的变形MR图像对进行配准。结果表明,与基于生物力学模型和基于学习的配准方法相比,所提出的方法达到了最先进的性能,并且在配准后成功地保证了软组织的生物力学约束。代码可在https://github.com/ZheMin-1992/Registration_PINNs上获得。
{"title":"Biomechanics-informed Non-rigid Medical Image Registration with Elasticity Theories.","authors":"Zhe Min, Zachary M C Baum, Shaheer U Saeed, Shixing Ma, Xinzhe Du, Mark Emberton, Dean C Barratt, Zeike A Taylor, Yipeng Hu","doi":"10.1109/TMI.2026.3665279","DOIUrl":"https://doi.org/10.1109/TMI.2026.3665279","url":null,"abstract":"<p><p>Biomechanical modelling of soft tissue provides a method for constraining medical image registration, such that the estimated spatial transformation is considered biophysically plausible. Existing methods either directly optimize the loss function containing the biomechanical-constrained regularization term over deformations, which takes much computational time, or are trained using biomechanically plausible data generated via finite element simulation, which is cumbersome. This work first instantiates the recently-proposed physics-informed neural networks (PINNs) to 3D elastic models that are used to establish the partial differential equations (PDEs) representing physics laws of biomechanical constraints to be satisfied. The registration algorithm that aligns point sets considering PINN-imposed biomechanics (i.e., the forward problem) is then formulated. In addition, the inverse problem and its algorithm of physical parameter (i.e., material property) estimation along with the registration are also formulated and developed. We carefully compare linear and nonlinear elasticity theories' capabilities in solving both tasks of forward registration and inverse physical parameter identification under PINNs respectively. Furthermore, two specific network configurations that leverage one common branch or two individual branches to predict deformation vectors and biomechanical states are also constructed and compared. The proposed PINNs-based registration approaches have been extensively evaluated with three experiments, that is single and multiple patient MRI-US registration using clinical MRI-US pairs, and registration using pairs of undeformed MR images from clinical cases of prostate cancer biopsy and deformed counterparts with finite-element-computed ground-truth deformation. Results demonstrate that the proposed methods achieve state-of-the-art performances compared to biomechanical-model-based and learning-based registration approaches, and the biomechanical constraints of soft tissues have been successfully warranted after registration. The codes are available at https://github.com/ZheMin-1992/Registration_PINNs.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147277926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MambaMatch: A Novel Model for Semi-Supervised Spinal Cortical and Cancellous Bone Segmentation. MambaMatch:一个半监督脊髓皮质骨和松质骨分割的新模型。
Pub Date : 2026-02-23 DOI: 10.1109/TMI.2026.3666480
Lianliang Li, Xue Li, Wenxin Chen, Sida Lyu, Changsheng Li, Xingguang Duan

Precise cortical and cancellous bone segmentation is essential for safe laminectomy. However, it remains challenging due to limited annotations and high anatomical similarity. To mitigate these challenges, we propose MambaMatch, an end-to-end semi-supervised segmentation framework collaboratively optimized under a teacher-student paradigm. The student branch adopts an innovative Mamba-ASPP-Unet (MAU) module, which integrates multi-scale Atrous Spatial Pyramid Pooling (ASPP) with channel and spatial attention to actively extract global spinal structures and cortical boundary features, while the teacher branch guides the student through constraints imposed by the KL-divergence. To enhance feature diversity, a dual-stream perturbation strategy is employed, combining Correlation-Guided CutMix Augmentation (CGCA) on high-response regions with standard strong augmentations. Furthermore, to account for inter-class variations in pseudo-label reliability, dynamic thresholding and temperature scaling are further introduced to adaptively balance pseudo-label selection and the intensity of consistency loss. MambaMatch achieves an mIoU of 79.69% and a Dice of 83.57% on our clinically validated CT Cortical- Cancellous Dataset, and also shows strong robustness on MRI-SPIDER, SKIN-ISIC 2018, PH2, and CT-VerSe datasets and the model remains lightweight and efficient. These results show that MambaMatch provides an efficient and accurate segmentation framework with clear potential to support clinical workflows in spinal surgery, while also demonstrating value across a broader range of clinical imaging scenarios.

准确的皮质骨和松质骨分割是安全椎板切除术的必要条件。然而,由于有限的注释和高度的解剖相似性,它仍然具有挑战性。为了缓解这些挑战,我们提出了MambaMatch,这是一个在师生模式下协同优化的端到端半监督分割框架。学生分支采用创新的Mamba-ASPP-Unet (MAU)模块,该模块将多尺度亚特拉斯空间金字塔池(ASPP)与通道和空间注意力相结合,主动提取全局脊柱结构和皮质边界特征,而教师分支则通过KL-divergence的约束引导学生。为了增强特征多样性,采用了双流微扰策略,将高响应区域的相关引导CutMix增强(CGCA)与标准强增强相结合。此外,为了考虑伪标签可靠性的类间变化,进一步引入了动态阈值和温度标度来自适应平衡伪标签选择和一致性损失的强度。MambaMatch在临床验证的CT皮质-松质数据集上实现了79.69%的mIoU和83.57%的Dice,并且在MRI-SPIDER、SKIN-ISIC 2018、PH2和CT- verse数据集上也表现出很强的鲁棒性,并且保持了轻量级和高效。这些结果表明,MambaMatch提供了一个高效、准确的分割框架,具有支持脊柱外科临床工作流程的明显潜力,同时也在更广泛的临床成像场景中展示了价值。
{"title":"MambaMatch: A Novel Model for Semi-Supervised Spinal Cortical and Cancellous Bone Segmentation.","authors":"Lianliang Li, Xue Li, Wenxin Chen, Sida Lyu, Changsheng Li, Xingguang Duan","doi":"10.1109/TMI.2026.3666480","DOIUrl":"https://doi.org/10.1109/TMI.2026.3666480","url":null,"abstract":"<p><p>Precise cortical and cancellous bone segmentation is essential for safe laminectomy. However, it remains challenging due to limited annotations and high anatomical similarity. To mitigate these challenges, we propose MambaMatch, an end-to-end semi-supervised segmentation framework collaboratively optimized under a teacher-student paradigm. The student branch adopts an innovative Mamba-ASPP-Unet (MAU) module, which integrates multi-scale Atrous Spatial Pyramid Pooling (ASPP) with channel and spatial attention to actively extract global spinal structures and cortical boundary features, while the teacher branch guides the student through constraints imposed by the KL-divergence. To enhance feature diversity, a dual-stream perturbation strategy is employed, combining Correlation-Guided CutMix Augmentation (CGCA) on high-response regions with standard strong augmentations. Furthermore, to account for inter-class variations in pseudo-label reliability, dynamic thresholding and temperature scaling are further introduced to adaptively balance pseudo-label selection and the intensity of consistency loss. MambaMatch achieves an mIoU of 79.69% and a Dice of 83.57% on our clinically validated CT Cortical- Cancellous Dataset, and also shows strong robustness on MRI-SPIDER, SKIN-ISIC 2018, PH<sup>2</sup>, and CT-VerSe datasets and the model remains lightweight and efficient. These results show that MambaMatch provides an efficient and accurate segmentation framework with clear potential to support clinical workflows in spinal surgery, while also demonstrating value across a broader range of clinical imaging scenarios.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147277981","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semantic Augmentation Variational Autoencoder for Unsupervised Anomaly Detection in Retinal OCT Images. 语义增强变分自编码器用于视网膜OCT图像的无监督异常检测。
Pub Date : 2026-02-23 DOI: 10.1109/TMI.2026.3667146
Xueying Zhou, Sijie Niu, Xiangmin Han, Xiaohui Li, Xizhan Gao, Guang Feng, Jun Shi

Performing unsupervised anomaly detection in retinal optical coherence tomography (OCT) images involves training a model solely on anomaly-free samples and detecting anomalies during inference, which reduces the cost of collecting large-scale annotated anomalous data. However, retinal OCT images exhibit significant variations in shape, thickness, and orientation, and lesions often have similar reflectance signals as normal tissues, making anomaly localization highly challenging. Existing methods address these challenges by flattening retinal layers, normalizing thickness, or leveraging reflectance priors, but their reliance on complex pre- and post-processing introduces uncertainties and limits end-to-end clinical applicability. To overcome these issues, we propose a novel semantic augmentation variational autoencoder (SeAugVAE) for unsupervised anomaly detection in retinal OCT images. Specifically, to capture the anatomical variability of normal retinas and thereby enhance anomaly sensitivity, we introduce a self-supervised semantic data augmentation strategy that enforces dual distribution consistency in both image and feature spaces during VAE training. For precise anomaly localization, we develop structural-semantic anomaly attention maps in the inference phase to detect anomalies from both local and global perspectives, and combine them to calculate anomaly score maps as the metric for localizing anomalous regions in images. Extensive experiments on multiple publicly and privately collected Cirrus and Spectralis OCT datasets demonstrate the effectiveness of SeAugVAE in pixel-wise unsupervised anomaly detection across multiple retinal diseases. Our codes are available at https://github.com/xyzhou1121/SeAugVAE.

在视网膜光学相干断层扫描(OCT)图像中进行无监督异常检测涉及到仅在无异常样本上训练模型并在推理过程中检测异常,从而降低了收集大规模带注释的异常数据的成本。然而,视网膜OCT图像在形状、厚度和方向上表现出明显的变化,病变通常具有与正常组织相似的反射信号,这使得异常定位非常具有挑战性。现有的方法通过使视网膜层变平、厚度归一化或利用反射先验来解决这些挑战,但它们对复杂的预处理和后处理的依赖带来了不确定性,限制了端到端的临床适用性。为了克服这些问题,我们提出了一种新的语义增强变分自编码器(SeAugVAE)用于视网膜OCT图像的无监督异常检测。具体来说,为了捕获正常视网膜的解剖变异,从而提高异常敏感性,我们引入了一种自监督语义数据增强策略,该策略在VAE训练期间在图像和特征空间中强制双重分布一致性。为了精确定位异常,我们在推理阶段开发了结构语义异常注意图,从局部和全局角度检测异常,并将它们结合起来计算异常评分图,作为图像中异常区域定位的度量。在多个公开和私人收集的Cirrus和Spectralis OCT数据集上进行的大量实验表明,SeAugVAE在跨多种视网膜疾病的逐像素无监督异常检测中是有效的。我们的代码可在https://github.com/xyzhou1121/SeAugVAE上获得。
{"title":"Semantic Augmentation Variational Autoencoder for Unsupervised Anomaly Detection in Retinal OCT Images.","authors":"Xueying Zhou, Sijie Niu, Xiangmin Han, Xiaohui Li, Xizhan Gao, Guang Feng, Jun Shi","doi":"10.1109/TMI.2026.3667146","DOIUrl":"https://doi.org/10.1109/TMI.2026.3667146","url":null,"abstract":"<p><p>Performing unsupervised anomaly detection in retinal optical coherence tomography (OCT) images involves training a model solely on anomaly-free samples and detecting anomalies during inference, which reduces the cost of collecting large-scale annotated anomalous data. However, retinal OCT images exhibit significant variations in shape, thickness, and orientation, and lesions often have similar reflectance signals as normal tissues, making anomaly localization highly challenging. Existing methods address these challenges by flattening retinal layers, normalizing thickness, or leveraging reflectance priors, but their reliance on complex pre- and post-processing introduces uncertainties and limits end-to-end clinical applicability. To overcome these issues, we propose a novel semantic augmentation variational autoencoder (SeAugVAE) for unsupervised anomaly detection in retinal OCT images. Specifically, to capture the anatomical variability of normal retinas and thereby enhance anomaly sensitivity, we introduce a self-supervised semantic data augmentation strategy that enforces dual distribution consistency in both image and feature spaces during VAE training. For precise anomaly localization, we develop structural-semantic anomaly attention maps in the inference phase to detect anomalies from both local and global perspectives, and combine them to calculate anomaly score maps as the metric for localizing anomalous regions in images. Extensive experiments on multiple publicly and privately collected Cirrus and Spectralis OCT datasets demonstrate the effectiveness of SeAugVAE in pixel-wise unsupervised anomaly detection across multiple retinal diseases. Our codes are available at https://github.com/xyzhou1121/SeAugVAE.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147277924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Universal Scale Transformer for Histology Image Segmentation. 用于组织图像分割的通用尺度转换器。
Pub Date : 2026-02-23 DOI: 10.1109/TMI.2026.3667069
Junjia Huang, Haofeng Li, Xiang Wan, Yuanhuan Xiong, Guanbin Li

Histology image segmentation is a critical prerequisite to pathological diagnosis. Accurate segmentation of these images can significantly aid physicians by facilitating quicker and more precise diagnostic decisions. A notable challenge in this area arises from the fact that different objects within histology images require segmentation at varying magnifications. However, most existing models are typically limited to performing segmentation at one pre-determined magnification. In this paper, we propose a novel universal scale transformer model (UniScaleFormer), which employs a scale-aware approach and integrates textual input to uniformly segment objects in histology images across various magnifications. Our method adopts an end-to-end architecture and utilizes a candidate mask query mechanism, specifically designed to identify and distinguish objects at different image scales. Moreover, we develop a Scale-Aware Module that enhances our network's ability to recognize the magnification level of input histology images, by using a scale query with extracted visual features. Then the scale query is integrated with mask queries, facilitating the incorporation of scale information. Experimental results demonstrate that the proposed method effectively achieves competitive results on various segmentation benchmarks at different magnifications. The code will be released at https://github.com/lhaof/UniScale.

组织图像分割是病理诊断的重要前提。这些图像的准确分割可以通过促进更快和更精确的诊断决策显着帮助医生。在这一领域的一个显著挑战是,组织学图像中的不同对象需要在不同的放大倍率下进行分割。然而,大多数现有模型通常仅限于在一个预先确定的放大倍率下执行分割。在本文中,我们提出了一种新的通用尺度转换器模型(UniScaleFormer),该模型采用尺度感知方法并集成文本输入,以在各种放大倍率下统一分割组织学图像中的对象。我们的方法采用端到端架构,利用候选掩码查询机制,专门用于识别和区分不同图像尺度下的目标。此外,我们开发了一个尺度感知模块,通过使用提取的视觉特征的尺度查询,增强了我们的网络识别输入组织学图像放大水平的能力。然后将尺度查询与掩码查询相结合,便于尺度信息的融合。实验结果表明,在不同的放大倍数下,该方法在不同的分割基准上取得了较好的分割效果。代码将在https://github.com/lhaof/UniScale上发布。
{"title":"Universal Scale Transformer for Histology Image Segmentation.","authors":"Junjia Huang, Haofeng Li, Xiang Wan, Yuanhuan Xiong, Guanbin Li","doi":"10.1109/TMI.2026.3667069","DOIUrl":"https://doi.org/10.1109/TMI.2026.3667069","url":null,"abstract":"<p><p>Histology image segmentation is a critical prerequisite to pathological diagnosis. Accurate segmentation of these images can significantly aid physicians by facilitating quicker and more precise diagnostic decisions. A notable challenge in this area arises from the fact that different objects within histology images require segmentation at varying magnifications. However, most existing models are typically limited to performing segmentation at one pre-determined magnification. In this paper, we propose a novel universal scale transformer model (UniScaleFormer), which employs a scale-aware approach and integrates textual input to uniformly segment objects in histology images across various magnifications. Our method adopts an end-to-end architecture and utilizes a candidate mask query mechanism, specifically designed to identify and distinguish objects at different image scales. Moreover, we develop a Scale-Aware Module that enhances our network's ability to recognize the magnification level of input histology images, by using a scale query with extracted visual features. Then the scale query is integrated with mask queries, facilitating the incorporation of scale information. Experimental results demonstrate that the proposed method effectively achieves competitive results on various segmentation benchmarks at different magnifications. The code will be released at https://github.com/lhaof/UniScale.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147277971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PWD: Prior-aware Wavelet Diffusion for Efficient Dental Limited-angle CT Reconstruction. 基于先验感知小波扩散的牙齿有限角CT高效重建。
Pub Date : 2026-02-23 DOI: 10.1109/TMI.2026.3667384
Yi Liu, Yiyang Wen, Zekun Zhou, Junqi Ma, Linghang Wang, Yucheng Yao, Liu Shi, Qiegen Liu

Generative diffusion models have received increasing attention in medical imaging, particularly in limited-angle computed tomography (LACT). Standard diffusion models achieve high-quality image reconstruction but require a large number of sampling steps during inference, imposing a heavy computational burden. Although skip-sampling strategies have been proposed to improve efficiency, they often lead to loss of fine structural details. To address this issue, we propose a Prior-aware Wavelet Diffusion for Efficient Dental Limited-angle CT Reconstruction (PWD). The PWD enables efficient sampling while preserving reconstruction fidelity in LACT, and effectively mitigates the degradation typically introduced by skip-sampling. Specifically, during the training phase, PWD maps the distribution of LACT images to that of fully sampled target images, enabling the model to learn structural correspondences between them. During inference, the LACT image serves as an prior-aware to guide the sampling trajectory, allowing for high-quality reconstruction with significantly fewer steps. In addition, PWD performs multi-scale feature fusion in the wavelet domain, effectively enhancing the reconstruction of fine details by leveraging both low-frequency and high-frequency information. Quantitative and qualitative evaluations on clinical dental arch CBCT and periapical datasets demonstrate that PWD outperforms existing methods under the same sampling condition. Using only 50 sampling steps, PWD achieves at least 1.7 dB improvement in PSNR and 10% gain in SSIM.

生成扩散模型在医学成像,特别是在有限角度计算机断层扫描(LACT)中受到越来越多的关注。标准的扩散模型可以实现高质量的图像重建,但在推理过程中需要大量的采样步骤,计算量很大。虽然跳过采样策略已被提出以提高效率,但它们往往导致精细结构细节的丢失。为了解决这个问题,我们提出了一种基于先验感知的小波扩散方法,用于牙齿有限角CT的高效重建。PWD能够在保持LACT重建保真度的同时实现高效采样,并有效减轻通常由跳过采样引入的退化。具体而言,在训练阶段,PWD将LACT图像的分布映射到完全采样的目标图像的分布,使模型能够学习它们之间的结构对应关系。在推理过程中,LACT图像作为先验感知来指导采样轨迹,允许以更少的步骤进行高质量的重建。此外,PWD在小波域进行多尺度特征融合,利用低频和高频信息有效增强了精细细节的重建。对临床牙弓CBCT和根尖周数据集的定量和定性评价表明,在相同采样条件下,PWD优于现有方法。仅使用50个采样步骤,PWD实现了至少1.7 dB的PSNR改进和10%的SSIM增益。
{"title":"PWD: Prior-aware Wavelet Diffusion for Efficient Dental Limited-angle CT Reconstruction.","authors":"Yi Liu, Yiyang Wen, Zekun Zhou, Junqi Ma, Linghang Wang, Yucheng Yao, Liu Shi, Qiegen Liu","doi":"10.1109/TMI.2026.3667384","DOIUrl":"https://doi.org/10.1109/TMI.2026.3667384","url":null,"abstract":"<p><p>Generative diffusion models have received increasing attention in medical imaging, particularly in limited-angle computed tomography (LACT). Standard diffusion models achieve high-quality image reconstruction but require a large number of sampling steps during inference, imposing a heavy computational burden. Although skip-sampling strategies have been proposed to improve efficiency, they often lead to loss of fine structural details. To address this issue, we propose a Prior-aware Wavelet Diffusion for Efficient Dental Limited-angle CT Reconstruction (PWD). The PWD enables efficient sampling while preserving reconstruction fidelity in LACT, and effectively mitigates the degradation typically introduced by skip-sampling. Specifically, during the training phase, PWD maps the distribution of LACT images to that of fully sampled target images, enabling the model to learn structural correspondences between them. During inference, the LACT image serves as an prior-aware to guide the sampling trajectory, allowing for high-quality reconstruction with significantly fewer steps. In addition, PWD performs multi-scale feature fusion in the wavelet domain, effectively enhancing the reconstruction of fine details by leveraging both low-frequency and high-frequency information. Quantitative and qualitative evaluations on clinical dental arch CBCT and periapical datasets demonstrate that PWD outperforms existing methods under the same sampling condition. Using only 50 sampling steps, PWD achieves at least 1.7 dB improvement in PSNR and 10% gain in SSIM.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147277909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FUGC: Benchmarking Semi-Supervised Learning Methods for Cervical Segmentation. 基于基准的半监督学习方法的颈部分割。
Pub Date : 2026-02-19 DOI: 10.1109/TMI.2026.3666364
Jieyun Bai, Yitong Tang, Zihao Zhou, Mahdi Islam, Musarrat Tabassum, Enrique Almar-Munoz, Hongyu Liu, Hui Meng, Nianjiang Lv, Bo Deng, Yu Chen, Zilun Peng, Yusong Xiao, Li Xiao, Nam-Khanh Tran, Dac-Phu Phan-Le, Hai-Dang Nguyen, Xiao Liu, Jiale Hu, Mingxu Huang, Jitao Liang, Chaolu Feng, Xuezhi Zhang, Lyuyang Tong, Bo Du, Ha-Hieu Pham, Thanh-Huy Nguyen, Min Xu, Juntao Jiang, Jiangning Zhang, Yong Liu, Md Kamrul Hasan, Jie Gan, Zhuonan Liang, Weidong Cai, Yuxin Huang, Gongning Luo, Mohammad Yaqub, Karim Lekadir

Accurate segmentation of cervical structures in transvaginal ultrasound (TVS) is critical for assessing the risk of spontaneous preterm birth (PTB), yet the scarcity of labeled data limits the performance of supervised learning approaches. This paper introduces the Fetal Ultrasound Grand Challenge (FUGC), the first benchmark for semi-supervised learning in cervical segmentation, hosted at ISBI 2025. FUGC provides a dataset of 890 TVS images, including 500 training images, 90 validation images, and 300 test images. Methods were evaluated using the Dice Similarity Coefficient (DSC), Hausdorff Distance (HD), and runtime (RT), with a weighted combination of 0.4/0.4/0.2. The challenge attracted 10 teams with 82 participants submitting innovative solutions. The best-performing methods for each individual metric achieved 90.26% mDSC, 38.88 mHD, and 32.85 ms RT, respectively. FUGC establishes a standardized benchmark for cervical segmentation, demonstrates the efficacy of semi-supervised methods with limited labeled data, and provides a foundation for AI-assisted clinical PTB risk assessment.

经阴道超声(TVS)中宫颈结构的准确分割对于评估自发性早产(PTB)的风险至关重要,但标记数据的稀缺性限制了监督学习方法的性能。本文介绍了胎儿超声大挑战(FUGC),这是ISBI 2025主办的宫颈分割半监督学习的第一个基准。FUGC提供了890个TVS图像的数据集,包括500个训练图像,90个验证图像和300个测试图像。采用Dice Similarity Coefficient (DSC)、Hausdorff Distance (HD)和runtime (RT)对各方法进行评价,加权组合为0.4/0.4/0.2。这项挑战吸引了10个团队,共有82名参与者提交了创新的解决方案。每个指标的最佳表现方法分别达到90.26% mDSC, 38.88 mHD和32.85 ms RT。FUGC建立了标准化的宫颈分割基准,在有限的标记数据下证明了半监督方法的有效性,为人工智能辅助临床PTB风险评估提供了基础。
{"title":"FUGC: Benchmarking Semi-Supervised Learning Methods for Cervical Segmentation.","authors":"Jieyun Bai, Yitong Tang, Zihao Zhou, Mahdi Islam, Musarrat Tabassum, Enrique Almar-Munoz, Hongyu Liu, Hui Meng, Nianjiang Lv, Bo Deng, Yu Chen, Zilun Peng, Yusong Xiao, Li Xiao, Nam-Khanh Tran, Dac-Phu Phan-Le, Hai-Dang Nguyen, Xiao Liu, Jiale Hu, Mingxu Huang, Jitao Liang, Chaolu Feng, Xuezhi Zhang, Lyuyang Tong, Bo Du, Ha-Hieu Pham, Thanh-Huy Nguyen, Min Xu, Juntao Jiang, Jiangning Zhang, Yong Liu, Md Kamrul Hasan, Jie Gan, Zhuonan Liang, Weidong Cai, Yuxin Huang, Gongning Luo, Mohammad Yaqub, Karim Lekadir","doi":"10.1109/TMI.2026.3666364","DOIUrl":"https://doi.org/10.1109/TMI.2026.3666364","url":null,"abstract":"<p><p>Accurate segmentation of cervical structures in transvaginal ultrasound (TVS) is critical for assessing the risk of spontaneous preterm birth (PTB), yet the scarcity of labeled data limits the performance of supervised learning approaches. This paper introduces the Fetal Ultrasound Grand Challenge (FUGC), the first benchmark for semi-supervised learning in cervical segmentation, hosted at ISBI 2025. FUGC provides a dataset of 890 TVS images, including 500 training images, 90 validation images, and 300 test images. Methods were evaluated using the Dice Similarity Coefficient (DSC), Hausdorff Distance (HD), and runtime (RT), with a weighted combination of 0.4/0.4/0.2. The challenge attracted 10 teams with 82 participants submitting innovative solutions. The best-performing methods for each individual metric achieved 90.26% mDSC, 38.88 mHD, and 32.85 ms RT, respectively. FUGC establishes a standardized benchmark for cervical segmentation, demonstrates the efficacy of semi-supervised methods with limited labeled data, and provides a foundation for AI-assisted clinical PTB risk assessment.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146230278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Axon Diameter Mapping from Myelin Water Diffusion MRI. 髓鞘水扩散MRI轴突直径成像。
Pub Date : 2026-02-13 DOI: 10.1109/TMI.2026.3664328
Hong-Hsi Lee, Kwok-Shing Chan, Dmitry S Novikov, Els Fieremans, Susie Y Huang

Probing diffusion in myelin water using diffusion-weighted T1-/T2-selective MRI acquisitions enables noninvasive measurement of myelinated axon diameter. Its application for in vivo measurements requires numerical verification through diffusion simulations. Here, we propose the theory of myelin water diffusion as measured with a diffusion MRI pulse sequence with wide gradient pulses using the Gaussian phase approximation. We establish its applicability to axonal diameter mapping via Monte Carlo simulations in either infinitely thin cylindrical surfaces or concentric cylindrical shells of finite thickness, mimicking the micro-geometry of myelin sheaths. The estimated diameters are shown to be weighted more toward outer than inner calibers. Simulation results evaluate the theory of myelin water diffusion and axon diameter estimation using spherical mean diffusion signals, demonstrating its applicability at signal-to-noise ratio above 20 on the Connectome 2.0 MRI scanner equipped with maximum gradient strength of 500 mT/m and slew rate of 600 T/m/s. Measuring restricted diffusion of myelin water in-between myelin sheaths using diffusion MRI allows one to measure myelinated axon diameters in vivo. The protocol can potentially be adapted for clinically available high-gradient performance scanners.

利用弥散加权T1 / t2选择性MRI采集探测髓鞘水中的扩散,可以无创测量髓鞘轴突直径。它在体内测量中的应用需要通过扩散模拟进行数值验证。在这里,我们提出了髓鞘水扩散的理论,通过使用高斯相位近似的宽梯度脉冲的扩散MRI脉冲序列来测量。我们通过蒙特卡罗模拟,在无限薄的圆柱形表面或有限厚度的同心圆柱形壳中,模拟髓鞘的微观几何结构,建立了它对轴索直径映射的适用性。估计直径的权重更倾向于外径而不是内径。模拟结果评估了髓鞘水扩散理论和利用球面平均扩散信号估计轴突直径的理论,证明了其在最大梯度强度为500 mT/m、旋转速率为600 T/m/s的Connectome 2.0 MRI扫描仪上信噪比大于20时的适用性。利用扩散MRI测量髓鞘间髓鞘水的受限扩散,可以在体内测量髓鞘轴突直径。该方案可以潜在地适用于临床可用的高梯度性能扫描仪。
{"title":"Axon Diameter Mapping from Myelin Water Diffusion MRI.","authors":"Hong-Hsi Lee, Kwok-Shing Chan, Dmitry S Novikov, Els Fieremans, Susie Y Huang","doi":"10.1109/TMI.2026.3664328","DOIUrl":"https://doi.org/10.1109/TMI.2026.3664328","url":null,"abstract":"<p><p>Probing diffusion in myelin water using diffusion-weighted T<sub>1</sub>-/T<sub>2</sub>-selective MRI acquisitions enables noninvasive measurement of myelinated axon diameter. Its application for in vivo measurements requires numerical verification through diffusion simulations. Here, we propose the theory of myelin water diffusion as measured with a diffusion MRI pulse sequence with wide gradient pulses using the Gaussian phase approximation. We establish its applicability to axonal diameter mapping via Monte Carlo simulations in either infinitely thin cylindrical surfaces or concentric cylindrical shells of finite thickness, mimicking the micro-geometry of myelin sheaths. The estimated diameters are shown to be weighted more toward outer than inner calibers. Simulation results evaluate the theory of myelin water diffusion and axon diameter estimation using spherical mean diffusion signals, demonstrating its applicability at signal-to-noise ratio above 20 on the Connectome 2.0 MRI scanner equipped with maximum gradient strength of 500 mT/m and slew rate of 600 T/m/s. Measuring restricted diffusion of myelin water in-between myelin sheaths using diffusion MRI allows one to measure myelinated axon diameters in vivo. The protocol can potentially be adapted for clinically available high-gradient performance scanners.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146196037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LRD-ESR-Net: Pseudo-healthy Image Synthesis Based on Low-resolution Residual Decoupling and Edge-prior-guided Super-resolution Reconstruction Module. LRD-ESR-Net:基于低分辨率残差解耦和边缘先验引导超分辨率重构模块的伪健康图像合成。
Pub Date : 2026-02-09 DOI: 10.1109/TMI.2026.3662706
Hang Gou, Wencong Zhang, Yujia Zhou, Qianjin Feng

Pseudo-healthy image synthesis aims to generate subject-specific, pathology-free images from pathological scans. Such images can be helpful in certain tasks, such as anomaly detection and understanding changes induced by pathology and disease. A participant cannot be "healthy" and "unhealthy" at the same time, and thus, directly obtaining pathological and healthy paired images of the same individual to train and evaluate supervised learning algorithms is infeasible. In addition, simultaneously meeting the requirements of subjects' "identity" preservation and pathology restoration performance is frequently difficult for existing unsupervised learning methods, especially for large or information-free pathological regions, such as postoperative cavities. In this study, we propose a novel pseudo-healthy synthesis framework that combines low-resolution residual decoupling with an edge-prior-guided super-resolution reconstruction module. We named this framework LRD-ESR-Net. In particular, by using a coarse-to-fine synthesis pipeline, the residual decoupling network first decouples information-rich tumor tissues or information-free resection cavities from healthy brain tissues in low-resolution pathological magnetic resonance images. Then, a residual-shifting diffusion network with Canny edge maps is employed to reconstruct low-resolution pseudo-healthy images to their original resolutions. We evaluate the proposed framework on one in-house brain dataset, two public brain datasets, and one public liver dataset, and validate its effectiveness on low-contrast lesion segmentation and pre-/postoperative brain tumor MRI registration. Results show that LRD-ESR-Net consistently outperforms state-of-the-art methods in pseudo-healthy image quality, anatomical preservation, and downstream task performance, demonstrating strong robustness and generalization across organs, modalities, and lesion types.

伪健康图像合成旨在从病理扫描中生成受试者特异性的无病理图像。这样的图像可以在某些任务中有所帮助,例如异常检测和理解由病理和疾病引起的变化。参与者不可能同时处于“健康”和“不健康”状态,因此,直接获取同一个体的病理和健康配对图像来训练和评估监督学习算法是不可行的。此外,现有的无监督学习方法往往难以同时满足受试者“身份”保存和病理恢复性能的要求,特别是对于较大或无信息的病理区域,如术后腔。在这项研究中,我们提出了一种新的伪健康综合框架,将低分辨率残差解耦与边缘先验引导的超分辨率重建模块相结合。我们将这个框架命名为LRD-ESR-Net。特别地,残差解耦网络利用粗到细的合成管道,首先将低分辨率病理磁共振图像中富含信息的肿瘤组织或无信息的切除腔与健康脑组织解耦。然后,利用残差移动扩散网络和Canny边缘图将低分辨率伪健康图像重建到原始分辨率;我们在一个内部大脑数据集、两个公共大脑数据集和一个公共肝脏数据集上评估了所提出的框架,并验证了其在低对比病灶分割和术前/术后脑肿瘤MRI配准方面的有效性。结果表明,LRD-ESR-Net在伪健康图像质量、解剖保存和下游任务性能方面始终优于最先进的方法,在器官、形态和病变类型方面表现出强大的鲁棒性和通用性。
{"title":"LRD-ESR-Net: Pseudo-healthy Image Synthesis Based on Low-resolution Residual Decoupling and Edge-prior-guided Super-resolution Reconstruction Module.","authors":"Hang Gou, Wencong Zhang, Yujia Zhou, Qianjin Feng","doi":"10.1109/TMI.2026.3662706","DOIUrl":"https://doi.org/10.1109/TMI.2026.3662706","url":null,"abstract":"<p><p>Pseudo-healthy image synthesis aims to generate subject-specific, pathology-free images from pathological scans. Such images can be helpful in certain tasks, such as anomaly detection and understanding changes induced by pathology and disease. A participant cannot be \"healthy\" and \"unhealthy\" at the same time, and thus, directly obtaining pathological and healthy paired images of the same individual to train and evaluate supervised learning algorithms is infeasible. In addition, simultaneously meeting the requirements of subjects' \"identity\" preservation and pathology restoration performance is frequently difficult for existing unsupervised learning methods, especially for large or information-free pathological regions, such as postoperative cavities. In this study, we propose a novel pseudo-healthy synthesis framework that combines low-resolution residual decoupling with an edge-prior-guided super-resolution reconstruction module. We named this framework LRD-ESR-Net. In particular, by using a coarse-to-fine synthesis pipeline, the residual decoupling network first decouples information-rich tumor tissues or information-free resection cavities from healthy brain tissues in low-resolution pathological magnetic resonance images. Then, a residual-shifting diffusion network with Canny edge maps is employed to reconstruct low-resolution pseudo-healthy images to their original resolutions. We evaluate the proposed framework on one in-house brain dataset, two public brain datasets, and one public liver dataset, and validate its effectiveness on low-contrast lesion segmentation and pre-/postoperative brain tumor MRI registration. Results show that LRD-ESR-Net consistently outperforms state-of-the-art methods in pseudo-healthy image quality, anatomical preservation, and downstream task performance, demonstrating strong robustness and generalization across organs, modalities, and lesion types.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146151529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Image Prior-Incorporated Direct PET Parametric Image Reconstruction. 深度图像先验融合直接PET参数图像重建。
Pub Date : 2026-02-09 DOI: 10.1109/TMI.2026.3662566
Andi Li, Mohammad B Syed, Jonathan B Moody, Jing Tang

Direct parametric reconstruction algorithms have been developed to improve the statistical reliability of parametric images estimated from dynamic PET imaging data. However, these estimates are degraded by noise due to measurement error and noise propagation during reconstruction. In this study, we develop a deep image prior (DIP) regularized direct reconstruction method, where the DIP network is used to represent the estimated parametric image. By initializing the DIP with pre-trained weights and updating its network to learn the intermediate information during reconstruction, the DIP regularization leverages the available population and subject-specific features. The proposed method is applied to reconstruct K1 from both simulated and patient data acquired by 82Rb dynamic PET myocardial perfusion imaging. Benefiting from the nonlinear representation capability of the DIP network, the proposed method achieves superior noise versus bias/mean performance compared with the indirect and direct reconstruction methods with various regularizations formed by quadratic smoothness, dictionary learning, or fully-connected neural network. To summarize, the proposed method demonstrates its potential in improving the precision of dynamic PET imaging measurements, which will contribute to diagnostic accuracy and disease monitoring.

为了提高动态PET成像数据估计的参数图像的统计可靠性,已经开发了直接参数重建算法。然而,由于测量误差和重构过程中的噪声传播,这些估计会受到噪声的影响。在本研究中,我们开发了一种深度图像先验(DIP)正则化直接重建方法,其中DIP网络用于表示估计的参数图像。通过使用预训练的权值初始化DIP并更新其网络以在重建过程中学习中间信息,DIP正则化利用了可用的总体和特定主题的特征。应用该方法从82Rb动态PET心肌灌注成像获得的模拟数据和患者数据中重建K1。利用DIP网络的非线性表示能力,与二次光滑、字典学习或全连接神经网络形成的各种正则化的间接和直接重建方法相比,该方法具有更好的噪声抗偏置/均值性能。总之,所提出的方法在提高动态PET成像测量精度方面具有潜力,这将有助于诊断准确性和疾病监测。
{"title":"Deep Image Prior-Incorporated Direct PET Parametric Image Reconstruction.","authors":"Andi Li, Mohammad B Syed, Jonathan B Moody, Jing Tang","doi":"10.1109/TMI.2026.3662566","DOIUrl":"10.1109/TMI.2026.3662566","url":null,"abstract":"<p><p>Direct parametric reconstruction algorithms have been developed to improve the statistical reliability of parametric images estimated from dynamic PET imaging data. However, these estimates are degraded by noise due to measurement error and noise propagation during reconstruction. In this study, we develop a deep image prior (DIP) regularized direct reconstruction method, where the DIP network is used to represent the estimated parametric image. By initializing the DIP with pre-trained weights and updating its network to learn the intermediate information during reconstruction, the DIP regularization leverages the available population and subject-specific features. The proposed method is applied to reconstruct K<sub>1</sub> from both simulated and patient data acquired by <sup>82</sup>Rb dynamic PET myocardial perfusion imaging. Benefiting from the nonlinear representation capability of the DIP network, the proposed method achieves superior noise versus bias/mean performance compared with the indirect and direct reconstruction methods with various regularizations formed by quadratic smoothness, dictionary learning, or fully-connected neural network. To summarize, the proposed method demonstrates its potential in improving the precision of dynamic PET imaging measurements, which will contribute to diagnostic accuracy and disease monitoring.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146151574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on medical imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1