首页 > 最新文献

Medical image analysis最新文献

英文 中文
AsymTrack: asymmetric fiber orientation mapping for accurate tractography in brain disorders via unsupervised deep learning AsymTrack:通过无监督深度学习对脑部疾病进行精确神经束造影的非对称纤维定向映射
IF 10.9 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-29 DOI: 10.1016/j.media.2026.103968
Di Zhang, Ziyu Li, Xiaofeng Deng, Zekun Han, Alan Wang, Yong Liu, Fangrong Zong
{"title":"AsymTrack: asymmetric fiber orientation mapping for accurate tractography in brain disorders via unsupervised deep learning","authors":"Di Zhang, Ziyu Li, Xiaofeng Deng, Zekun Han, Alan Wang, Yong Liu, Fangrong Zong","doi":"10.1016/j.media.2026.103968","DOIUrl":"https://doi.org/10.1016/j.media.2026.103968","url":null,"abstract":"","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"50 1","pages":""},"PeriodicalIF":10.9,"publicationDate":"2026-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146071491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A navigation-guided 3D breast ultrasound scanning and reconstruction system for automated multi-lesion spatial localization and diagnosis 一种用于多病灶自动定位与诊断的导航引导三维乳腺超声扫描与重建系统
IF 10.9 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-28 DOI: 10.1016/j.media.2026.103965
Yi Zhang, Yulin Yan, Kun Wang, Muyu Cai, Yifei Xiang, Yan Guo, Puxun Tu, Tao Ying, Xiaojun Chen
{"title":"A navigation-guided 3D breast ultrasound scanning and reconstruction system for automated multi-lesion spatial localization and diagnosis","authors":"Yi Zhang, Yulin Yan, Kun Wang, Muyu Cai, Yifei Xiang, Yan Guo, Puxun Tu, Tao Ying, Xiaojun Chen","doi":"10.1016/j.media.2026.103965","DOIUrl":"https://doi.org/10.1016/j.media.2026.103965","url":null,"abstract":"","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"76 1","pages":"103965"},"PeriodicalIF":10.9,"publicationDate":"2026-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146072194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Sparse Fusion Transformer Network with Spatio-Temporal Decoupling for Breast Tumor Classification 时空解耦的多模态稀疏融合变压器网络用于乳腺肿瘤分类
IF 10.9 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-28 DOI: 10.1016/j.media.2026.103966
Jiahao Xu, Shuxin Zhuang, Yi He, Haolin Wang, Zhemin Zhuang, Huancheng Zeng
{"title":"Multimodal Sparse Fusion Transformer Network with Spatio-Temporal Decoupling for Breast Tumor Classification","authors":"Jiahao Xu, Shuxin Zhuang, Yi He, Haolin Wang, Zhemin Zhuang, Huancheng Zeng","doi":"10.1016/j.media.2026.103966","DOIUrl":"https://doi.org/10.1016/j.media.2026.103966","url":null,"abstract":"","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"74 1","pages":""},"PeriodicalIF":10.9,"publicationDate":"2026-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146072192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Revisiting Lesion Tracking in 3D Total Body Photography 重新审视三维全身摄影中的病灶跟踪
IF 10.9 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-28 DOI: 10.1016/j.media.2026.103963
Wei-Lun Huang, Minghao Xue, Zhiyou Liu, Davood Tashayyod, Jun Kang, Amir Gandjbakhche, Misha Kazhdan, Mehran Armand
{"title":"Revisiting Lesion Tracking in 3D Total Body Photography","authors":"Wei-Lun Huang, Minghao Xue, Zhiyou Liu, Davood Tashayyod, Jun Kang, Amir Gandjbakhche, Misha Kazhdan, Mehran Armand","doi":"10.1016/j.media.2026.103963","DOIUrl":"https://doi.org/10.1016/j.media.2026.103963","url":null,"abstract":"","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"92 1","pages":""},"PeriodicalIF":10.9,"publicationDate":"2026-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146072193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DPFR: Semi-supervised gland segmentation via density perturbation and feature recalibration DPFR:通过密度扰动和特征重新校准的半监督腺体分割
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-27 DOI: 10.1016/j.media.2026.103962
Jiejiang Yu, Yu Liu
In recent years, semi-supervised methods have attracted considerable attention in gland segmentation of histopathological images, as they can substantially reduce the annotation data burden for pathologists. The most widely adopted approach is the Mean-Teacher framework based on consistency regularization, which exploits unlabeled data information through consistency regularization constraints. However, due to the morphological complexity of glands in histopathological images, existing methods still suffer from confusion between glands and background, as well as gland adhesion. To address these challenges, we propose a semi-supervised gland segmentation method based on Density Perturbation and Feature Recalibration (DPFR). Specifically, we first design a normalized flow-based density estimator to effectively model the feature density distributions of glands, contours, and background. The gradient information of the estimator is then exploited to determine the descent direction in low-density regions, along which perturbations are applied to enhance feature discriminability. Furthermore, a contrastive-learning-based feature recalibration module is designed to alleviate inter-class distribution confusion, thereby improving gland-background separability and mitigating gland adhesion. Extensive experiments on three public gland segmentation datasets demonstrate that the proposed method consistently outperforms existing semi-supervised approaches, achieving state-of-the-art performance with a substantial margin. The code repository address is https://github.com/Methow0/DPFR.
近年来,半监督方法在组织病理图像的腺体分割中引起了广泛的关注,因为它可以大大减少病理学家的注释数据负担。采用最广泛的方法是基于一致性正则化的Mean-Teacher框架,该框架通过一致性正则化约束来利用未标记的数据信息。然而,由于组织病理图像中腺体形态的复杂性,现有方法存在腺体与背景混淆、腺体粘连等问题。为了解决这些问题,我们提出了一种基于密度扰动和特征再校准(DPFR)的半监督腺体分割方法。具体来说,我们首先设计了一个归一化的基于流的密度估计器,以有效地模拟腺体、轮廓和背景的特征密度分布。然后利用估计器的梯度信息确定低密度区域的下降方向,沿此方向施加扰动以增强特征的可判别性。此外,设计了基于对比学习的特征再校准模块,以缓解类间分布混乱,从而提高腺体-背景可分离性,减轻腺体粘附。在三个公共腺体分割数据集上进行的大量实验表明,所提出的方法始终优于现有的半监督方法,在很大程度上实现了最先进的性能。代码存储库地址是https://github.com/Methow0/DPFR。
{"title":"DPFR: Semi-supervised gland segmentation via density perturbation and feature recalibration","authors":"Jiejiang Yu,&nbsp;Yu Liu","doi":"10.1016/j.media.2026.103962","DOIUrl":"10.1016/j.media.2026.103962","url":null,"abstract":"<div><div>In recent years, semi-supervised methods have attracted considerable attention in gland segmentation of histopathological images, as they can substantially reduce the annotation data burden for pathologists. The most widely adopted approach is the Mean-Teacher framework based on consistency regularization, which exploits unlabeled data information through consistency regularization constraints. However, due to the morphological complexity of glands in histopathological images, existing methods still suffer from confusion between glands and background, as well as gland adhesion. To address these challenges, we propose a semi-supervised gland segmentation method based on Density Perturbation and Feature Recalibration (DPFR). Specifically, we first design a normalized flow-based density estimator to effectively model the feature density distributions of glands, contours, and background. The gradient information of the estimator is then exploited to determine the descent direction in low-density regions, along which perturbations are applied to enhance feature discriminability. Furthermore, a contrastive-learning-based feature recalibration module is designed to alleviate inter-class distribution confusion, thereby improving gland-background separability and mitigating gland adhesion. Extensive experiments on three public gland segmentation datasets demonstrate that the proposed method consistently outperforms existing semi-supervised approaches, achieving state-of-the-art performance with a substantial margin. The code repository address is <span><span>https://github.com/Methow0/DPFR</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103962"},"PeriodicalIF":11.8,"publicationDate":"2026-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146072196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reliable uncertainty quantification for 2D/3D anatomical landmark localization using multi-output conformal prediction 使用多输出适形预测进行二维/三维解剖地标定位的可靠不确定性量化
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-27 DOI: 10.1016/j.media.2026.103953
Jef Jonkers , Frank Coopman , Luc Duchateau , Glenn Van Wallendael , Sofie Van Hoecke
Automatic anatomical landmark localization in medical imaging requires not just accurate predictions but reliable uncertainty quantification for effective clinical decision support. Current uncertainty quantification approaches often fall short, particularly when combined with normality assumptions, systematically underestimating total predictive uncertainty. This paper introduces conformal prediction as a framework for reliable uncertainty quantification in anatomical landmark localization, addressing a critical gap in automatic landmark localization. We present two novel approaches guaranteeing finite-sample validity for multi-output prediction: multi-output regression-as-classification conformal prediction (M-R2CCP) and its variant multi-output regression to classification conformal prediction set to region (M-R2C2R). Unlike conventional methods that produce axis-aligned hyperrectangular or ellipsoidal regions, our approaches generate flexible, non-convex prediction regions that better capture the underlying uncertainty structure of landmark predictions. Through extensive empirical evaluation across multiple 2D and 3D datasets, we demonstrate that our methods consistently outperform existing multi-output conformal prediction approaches in both validity and efficiency. This work represents a significant advancement in reliable uncertainty estimation for anatomical landmark localization, providing clinicians with trustworthy confidence measures for their diagnoses. While developed for medical imaging, these methods show promise for broader applications in multi-output regression problems.
医学成像中的自动解剖地标定位不仅需要准确的预测,还需要可靠的不确定性量化,以提供有效的临床决策支持。当前的不确定性量化方法往往不足,特别是当与正态性假设相结合时,系统地低估了总的预测不确定性。本文引入保形预测作为解剖学地标定位中可靠的不确定性量化框架,解决了自动地标定位的关键空白。提出了两种保证多输出预测有限样本有效性的新方法:多输出回归作为分类适形预测(M-R2CCP)及其变体多输出回归到分类适形预测集到区域(M-R2C2R)。与产生轴向超矩形或椭球体区域的传统方法不同,我们的方法产生灵活的非凸预测区域,可以更好地捕获里程碑预测的潜在不确定性结构。通过对多个2D和3D数据集的广泛经验评估,我们证明了我们的方法在有效性和效率方面始终优于现有的多输出适形预测方法。这项工作代表了解剖地标定位可靠不确定性估计的重大进展,为临床医生的诊断提供了值得信赖的信心措施。虽然这些方法是为医学成像而开发的,但它们在多输出回归问题中有更广泛的应用前景。
{"title":"Reliable uncertainty quantification for 2D/3D anatomical landmark localization using multi-output conformal prediction","authors":"Jef Jonkers ,&nbsp;Frank Coopman ,&nbsp;Luc Duchateau ,&nbsp;Glenn Van Wallendael ,&nbsp;Sofie Van Hoecke","doi":"10.1016/j.media.2026.103953","DOIUrl":"10.1016/j.media.2026.103953","url":null,"abstract":"<div><div>Automatic anatomical landmark localization in medical imaging requires not just accurate predictions but reliable uncertainty quantification for effective clinical decision support. Current uncertainty quantification approaches often fall short, particularly when combined with normality assumptions, systematically underestimating total predictive uncertainty. This paper introduces conformal prediction as a framework for reliable uncertainty quantification in anatomical landmark localization, addressing a critical gap in automatic landmark localization. We present two novel approaches guaranteeing finite-sample validity for multi-output prediction: multi-output regression-as-classification conformal prediction (M-R2CCP) and its variant multi-output regression to classification conformal prediction set to region (M-R2C2R). Unlike conventional methods that produce axis-aligned hyperrectangular or ellipsoidal regions, our approaches generate flexible, non-convex prediction regions that better capture the underlying uncertainty structure of landmark predictions. Through extensive empirical evaluation across multiple 2D and 3D datasets, we demonstrate that our methods consistently outperform existing multi-output conformal prediction approaches in both validity and efficiency. This work represents a significant advancement in reliable uncertainty estimation for anatomical landmark localization, providing clinicians with trustworthy confidence measures for their diagnoses. While developed for medical imaging, these methods show promise for broader applications in multi-output regression problems.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103953"},"PeriodicalIF":11.8,"publicationDate":"2026-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146071492","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MIL-Adapter: Coupling multiple instance learning and vision-language adapters for few-shot slide-level classification MIL-Adapter:耦合多实例学习和视觉语言适配器,用于少量幻灯片级别的分类
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-26 DOI: 10.1016/j.media.2026.103964
Pablo Meseguer , Rocío del Amor , Valery Naranjo
Contrastive language-image pretraining has greatly enhanced visual representation learning and enabled zero-shot classification. Vision-language language models (VLM) have succeeded in few-shot learning by leveraging adaptation modules fine-tuned for specific downstream tasks. In computational pathology (CPath), accurate whole-slide image (WSI) prediction is crucial for aiding in cancer diagnosis, and multiple instance learning (MIL) remains essential for managing the gigapixel scale of WSIs. In the intersection between CPath and VLMs, the literature still lacks specific adapters that handle the particular complexity of the slides. To solve this gap, we introduce MIL-Adapter, a novel approach designed to obtain consistent slide-level classification under few-shot learning scenarios. In particular, our framework is the first to combine trainable MIL aggregation functions and lightweight visual-language adapters to improve the performance of histopathological VLMs. MIL-Adapter relies on textual ensemble learning to construct discriminative zero-shot prototypes. It is serves as a solid starting point, surpassing MIL models with randomly initialized classifiers in data-constrained settings. With our experimentation, we demonstrate the value of textual ensemble learning and the robust predictive performance of MIL-Adapter through diverse datasets and configurations of few-shot scenarios, while providing crucial insights on model interpretability. The code is publicly accessible in https://github.com/cvblab/MIL-Adapter.
对比语言图像预训练极大地增强了视觉表征学习,实现了零采样分类。视觉语言模型(VLM)通过利用针对特定下游任务进行微调的自适应模块,在少量学习中取得了成功。在计算病理学(CPath)中,准确的全片图像(WSI)预测对于帮助癌症诊断至关重要,而多实例学习(MIL)对于管理WSI的十亿像素规模仍然至关重要。在CPath和vlm之间的交集中,文献仍然缺乏处理幻灯片特定复杂性的特定适配器。为了解决这一差距,我们引入了MIL-Adapter,这是一种新的方法,旨在在少镜头学习场景下获得一致的幻灯片级别分类。特别是,我们的框架是第一个将可训练的MIL聚合功能和轻量级的视觉语言适配器结合起来以提高组织病理学vlm的性能的框架。MIL-Adapter依赖于文本集成学习来构建判别零射击原型。它可以作为一个坚实的起点,在数据约束设置中超越具有随机初始化分类器的MIL模型。通过实验,我们展示了文本集成学习的价值以及MIL-Adapter通过不同数据集和少量场景配置的鲁棒预测性能,同时提供了关于模型可解释性的重要见解。该代码可在https://github.com/cvblab/MIL-Adapter上公开访问。
{"title":"MIL-Adapter: Coupling multiple instance learning and vision-language adapters for few-shot slide-level classification","authors":"Pablo Meseguer ,&nbsp;Rocío del Amor ,&nbsp;Valery Naranjo","doi":"10.1016/j.media.2026.103964","DOIUrl":"10.1016/j.media.2026.103964","url":null,"abstract":"<div><div>Contrastive language-image pretraining has greatly enhanced visual representation learning and enabled zero-shot classification. Vision-language language models (VLM) have succeeded in few-shot learning by leveraging adaptation modules fine-tuned for specific downstream tasks. In computational pathology (CPath), accurate whole-slide image (WSI) prediction is crucial for aiding in cancer diagnosis, and multiple instance learning (MIL) remains essential for managing the gigapixel scale of WSIs. In the intersection between CPath and VLMs, the literature still lacks specific adapters that handle the particular complexity of the slides. To solve this gap, we introduce MIL-Adapter, a novel approach designed to obtain consistent slide-level classification under few-shot learning scenarios. In particular, our framework is the first to combine trainable MIL aggregation functions and lightweight visual-language adapters to improve the performance of histopathological VLMs. MIL-Adapter relies on textual ensemble learning to construct discriminative zero-shot prototypes. It is serves as a solid starting point, surpassing MIL models with randomly initialized classifiers in data-constrained settings. With our experimentation, we demonstrate the value of textual ensemble learning and the robust predictive performance of MIL-Adapter through diverse datasets and configurations of few-shot scenarios, while providing crucial insights on model interpretability. The code is publicly accessible in <span><span>https://github.com/cvblab/MIL-Adapter</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103964"},"PeriodicalIF":11.8,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146048254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Boundary Confusion for Volumetric Medical Image Segmentation 体医学图像分割中边界混淆的研究
IF 10.9 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-25 DOI: 10.1016/j.media.2026.103961
Xin You, Ming Ding, Minghui Zhang, Hanxiao Zhang, Junyang Wu, Yi Yu, Jie Yang, Yun Gu
Accurate boundary segmentation of volumetric images is a critical task for image-guided diagnosis and computer-assisted intervention. It is challenging to address the boundary confusion with explicit constraints. Existing methods of refining boundaries overemphasize the slender structure while overlooking the dynamic interactions between boundaries and neighboring regions. In this paper, we reconceptualize the mechanism of boundary generation via introducing Pushing and Pulling interactions, then propose a unified network termed PP-Net to model shape characteristics of the confused boundary region. Specifically, we first propose the semantic difference module (SDM) from the pushing branch to drive the boundary towards the ground truth under diffusion guidance. Additionally, the class clustering module (CCM) from the pulling branch is introduced to stretch the intersected boundary along the opposite direction. Thus, pushing and pulling branches will furnish two adversarial forces to enhance representation capabilities for the faint boundary. Experiments are conducted on four public datasets and one in-house dataset plagued by boundary confusion. The results demonstrate the superiority of PP-Net over other segmentation networks, especially on the evaluation metrics of Hausdorff Distance and Average Symmetric Surface Distance. Besides, SDM and CCM can serve as plug-and-play modules to enhance classic U-shape baseline models, including recent SAM-based foundation models. Source codes are available at https://github.com/EndoluminalSurgicalVision-IMR/PnPNet.
体积图像的精确边界分割是图像引导诊断和计算机辅助干预的关键任务。用明确的约束来解决边界混淆是具有挑战性的。现有的边界细化方法过分强调细长的结构,而忽略了边界与邻近区域之间的动态相互作用。本文通过引入推拉相互作用,重新定义了边界生成的机制,并提出了一个统一的PP-Net网络来模拟混乱边界区域的形状特征。具体而言,我们首先提出了推枝的语义差分模块(SDM),在扩散引导下将边界向地面真值驱动。此外,引入了来自拉分支的类聚类模块(CCM),将相交边界沿相反方向拉伸。因此,推和拉分支将提供两种对立的力量,以增强对模糊边界的表示能力。在四个公共数据集和一个边界混乱的内部数据集上进行了实验。结果表明,PP-Net在Hausdorff距离和平均对称表面距离的评价指标上优于其他分割网络。此外,SDM和CCM可以作为即插即用模块来增强经典的u型基线模型,包括最近基于sam的基础模型。源代码可从https://github.com/EndoluminalSurgicalVision-IMR/PnPNet获得。
{"title":"Towards Boundary Confusion for Volumetric Medical Image Segmentation","authors":"Xin You, Ming Ding, Minghui Zhang, Hanxiao Zhang, Junyang Wu, Yi Yu, Jie Yang, Yun Gu","doi":"10.1016/j.media.2026.103961","DOIUrl":"https://doi.org/10.1016/j.media.2026.103961","url":null,"abstract":"Accurate boundary segmentation of volumetric images is a critical task for image-guided diagnosis and computer-assisted intervention. It is challenging to address the boundary confusion with explicit constraints. Existing methods of refining boundaries overemphasize the slender structure while overlooking the dynamic interactions between boundaries and neighboring regions. In this paper, we reconceptualize the mechanism of boundary generation via introducing Pushing and Pulling interactions, then propose a unified network termed PP-Net to model shape characteristics of the confused boundary region. Specifically, we first propose the semantic difference module (SDM) from the pushing branch to drive the boundary towards the ground truth under diffusion guidance. Additionally, the class clustering module (CCM) from the pulling branch is introduced to stretch the intersected boundary along the opposite direction. Thus, pushing and pulling branches will furnish two adversarial forces to enhance representation capabilities for the faint boundary. Experiments are conducted on four public datasets and one in-house dataset plagued by boundary confusion. The results demonstrate the superiority of PP-Net over other segmentation networks, especially on the evaluation metrics of Hausdorff Distance and Average Symmetric Surface Distance. Besides, SDM and CCM can serve as plug-and-play modules to enhance classic U-shape baseline models, including recent SAM-based foundation models. Source codes are available at <ce:inter-ref xlink:href=\"https://github.com/EndoluminalSurgicalVision-IMR/PnPNet\" xlink:type=\"simple\">https://github.com/EndoluminalSurgicalVision-IMR/PnPNet</ce:inter-ref>.","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"291 1","pages":""},"PeriodicalIF":10.9,"publicationDate":"2026-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146048365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fundus image quality assessment in retinopathy of prematurity via multi-label graph evidential network 基于多标签图证据网络的早产儿视网膜病变眼底图像质量评价
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-24 DOI: 10.1016/j.media.2026.103959
Donghan Wu , Wenyue Shen , Lu Yuan , Heng Li , Huaying Hao , Juan Ye , Yitian Zhao
Retinopathy of Prematurity (ROP) is a leading cause of childhood blindness worldwide. In clinical practice, fundus imaging serves as a primary diagnostic tool for ROP, making the accurate quality assessment of these images critically important. However, existing automated methods for evaluating ROP fundus images face significant challenges. First, there is a high degree of visual similarity between lesions and factors that influence quality. Second, there is a paucity of trustworthy outputs and interpretable or clinical-friendly designs, which limit their reliability and effectiveness. In this work, we propose a ROP image quality assessment framework, termed Q-ROP. This framework leverages fine-grained multi-label annotations based on key image factors such as artifacts, illumination, spatial positioning, and structural clarity. Additionally, the integration of a label graph network with evidential learning theory enables the model to explicitly capture the relationships between quality grades and influencing factors, thereby improving both robustness and accuracy. This approach facilitates interpretable analysis by directing the model’s focus toward relevant image features and reducing interference from lesion-like artifacts. Furthermore, the incorporation of evidential learning theory serves to quantify the uncertainty inherent in quality ratings, thereby ensuring the trustworthiness of the assessments. Trained and tested on a dataset of 6677 ROP images across three quality levels (i.e. acceptable, potentially acceptable, and unacceptable), Q-ROP achieved state-of-the-art performance with a 95.82% accuracy. Its effectiveness was further validated in a downstream ROP staging task, where it significantly improved the performance of typical classification models. These results demonstrate Q-ROP’s strong potential as a reliable and robust tool for clinical decision support.
早产儿视网膜病变(ROP)是全球儿童失明的主要原因。在临床实践中,眼底成像是ROP的主要诊断工具,因此对这些图像的准确质量评估至关重要。然而,现有的眼底图像ROP评估自动化方法面临着重大挑战。首先,在病变和影响质量的因素之间存在高度的视觉相似性。其次,缺乏可信赖的输出和可解释或临床友好的设计,这限制了它们的可靠性和有效性。在这项工作中,我们提出了一个ROP图像质量评估框架,称为Q-ROP。该框架利用基于关键图像因素(如工件、照明、空间定位和结构清晰度)的细粒度多标签注释。此外,将标签图网络与证据学习理论相结合,使模型能够明确地捕捉质量等级与影响因素之间的关系,从而提高鲁棒性和准确性。这种方法通过将模型的焦点指向相关的图像特征并减少来自类似病变的工件的干扰,从而促进了可解释的分析。此外,证据学习理论的结合有助于量化质量评级固有的不确定性,从而确保评估的可信度。在三个质量水平(即可接受、潜在可接受和不可接受)的6677张ROP图像数据集上进行训练和测试,Q-ROP达到了最先进的性能,准确率为95.82%。在下游ROP分期任务中进一步验证了其有效性,该方法显著提高了典型分类模型的性能。这些结果表明Q-ROP作为临床决策支持的可靠和强大的工具具有强大的潜力。
{"title":"Fundus image quality assessment in retinopathy of prematurity via multi-label graph evidential network","authors":"Donghan Wu ,&nbsp;Wenyue Shen ,&nbsp;Lu Yuan ,&nbsp;Heng Li ,&nbsp;Huaying Hao ,&nbsp;Juan Ye ,&nbsp;Yitian Zhao","doi":"10.1016/j.media.2026.103959","DOIUrl":"10.1016/j.media.2026.103959","url":null,"abstract":"<div><div>Retinopathy of Prematurity (ROP) is a leading cause of childhood blindness worldwide. In clinical practice, fundus imaging serves as a primary diagnostic tool for ROP, making the accurate quality assessment of these images critically important. However, existing automated methods for evaluating ROP fundus images face significant challenges. First, there is a high degree of visual similarity between lesions and factors that influence quality. Second, there is a paucity of trustworthy outputs and interpretable or clinical-friendly designs, which limit their reliability and effectiveness. In this work, we propose a ROP image quality assessment framework, termed Q-ROP. This framework leverages fine-grained multi-label annotations based on key image factors such as artifacts, illumination, spatial positioning, and structural clarity. Additionally, the integration of a label graph network with evidential learning theory enables the model to explicitly capture the relationships between quality grades and influencing factors, thereby improving both robustness and accuracy. This approach facilitates interpretable analysis by directing the model’s focus toward relevant image features and reducing interference from lesion-like artifacts. Furthermore, the incorporation of evidential learning theory serves to quantify the uncertainty inherent in quality ratings, thereby ensuring the trustworthiness of the assessments. Trained and tested on a dataset of 6677 ROP images across three quality levels (i.e. acceptable, potentially acceptable, and unacceptable), Q-ROP achieved state-of-the-art performance with a 95.82% accuracy. Its effectiveness was further validated in a downstream ROP staging task, where it significantly improved the performance of typical classification models. These results demonstrate Q-ROP’s strong potential as a reliable and robust tool for clinical decision support.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103959"},"PeriodicalIF":11.8,"publicationDate":"2026-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146048255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fréchet radiomic distance (FRD): A versatile metric for comparing medical imaging datasets ferccheradiomic Distance (FRD):一种比较医学影像数据集的通用度量
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-24 DOI: 10.1016/j.media.2026.103943
Nicholas Konz , Richard Osuala , Preeti Verma , Yuwen Chen , Hanxue Gu , Haoyu Dong , Yaqian Chen , Andrew Marshall , Lidia Garrucho , Kaisar Kushibar , Daniel M. Lang , Gene S. Kim , Lars J. Grimm , John M. Lewin , James S. Duncan , Julia A. Schnabel , Oliver Diaz , Karim Lekadir , Maciej A. Mazurowski
Determining whether two sets of images belong to the same or different distributions or domains is a crucial task in modern medical image analysis and deep learning; for example, to evaluate the output quality of image generative models. Currently, metrics used for this task either rely on the (potentially biased) choice of some downstream task, such as segmentation, or adopt task-independent perceptual metrics (e.g., Fréchet Inception Distance/FID) from natural imaging, which we show insufficiently capture anatomical features. To this end, we introduce a new perceptual metric tailored for medical images, FRD (Fréchet Radiomic Distance), which utilizes standardized, clinically meaningful, and interpretable image features. We show that FRD is superior to other image distribution metrics for a range of medical imaging applications, including out-of-domain (OOD) detection, the evaluation of image-to-image translation (by correlating more with downstream task performance as well as anatomical consistency and realism), and the evaluation of unconditional image generation. Moreover, FRD offers additional benefits such as stability and computational efficiency at low sample sizes, sensitivity to image corruptions and adversarial attacks, feature interpretability, and correlation with radiologist-perceived image quality. Additionally, we address key gaps in the literature by presenting an extensive framework for the multifaceted evaluation of image similarity metrics in medical imaging—including the first large-scale comparative study of generative models for medical image translation—and release an accessible codebase to facilitate future research. Our results are supported by thorough experiments spanning a variety of datasets, modalities, and downstream tasks, highlighting the broad potential of FRD for medical image analysis.
确定两组图像是否属于相同或不同的分布或域是现代医学图像分析和深度学习的关键任务;例如,评估图像生成模型的输出质量。目前,用于该任务的度量要么依赖于(可能有偏见的)一些下游任务的选择,例如分割,要么采用与任务无关的感知度量(例如,fr起始距离/FID)来自自然成像,我们认为这不足以捕获解剖特征。为此,我们引入了一种为医学图像量身定制的新的感知度量,FRD (fr放射距离),它利用标准化、临床意义和可解释的图像特征。研究表明,在一系列医学成像应用中,FRD优于其他图像分布指标,包括域外(OOD)检测、图像到图像转换的评估(通过更多地与下游任务性能以及解剖一致性和真实感相关)以及无条件图像生成的评估。此外,FRD还提供了额外的好处,如低样本量下的稳定性和计算效率、对图像损坏和对抗性攻击的敏感性、特征可解释性以及与放射科医生感知的图像质量的相关性。此外,我们通过提出医学成像中图像相似性度量的多方面评估的广泛框架(包括医学图像翻译生成模型的首次大规模比较研究)来解决文献中的关键空白,并发布了一个可访问的代码库,以促进未来的研究。我们的结果得到了跨越各种数据集、模式和下游任务的全面实验的支持,突出了FRD在医学图像分析方面的广泛潜力。
{"title":"Fréchet radiomic distance (FRD): A versatile metric for comparing medical imaging datasets","authors":"Nicholas Konz ,&nbsp;Richard Osuala ,&nbsp;Preeti Verma ,&nbsp;Yuwen Chen ,&nbsp;Hanxue Gu ,&nbsp;Haoyu Dong ,&nbsp;Yaqian Chen ,&nbsp;Andrew Marshall ,&nbsp;Lidia Garrucho ,&nbsp;Kaisar Kushibar ,&nbsp;Daniel M. Lang ,&nbsp;Gene S. Kim ,&nbsp;Lars J. Grimm ,&nbsp;John M. Lewin ,&nbsp;James S. Duncan ,&nbsp;Julia A. Schnabel ,&nbsp;Oliver Diaz ,&nbsp;Karim Lekadir ,&nbsp;Maciej A. Mazurowski","doi":"10.1016/j.media.2026.103943","DOIUrl":"10.1016/j.media.2026.103943","url":null,"abstract":"<div><div>Determining whether two sets of images belong to the same or different distributions or domains is a crucial task in modern medical image analysis and deep learning; for example, to evaluate the output quality of image generative models. Currently, metrics used for this task either rely on the (potentially biased) choice of some downstream task, such as segmentation, or adopt task-independent perceptual metrics (<em>e.g.</em>, Fréchet Inception Distance/FID) from natural imaging, which we show insufficiently capture anatomical features. To this end, we introduce a new perceptual metric tailored for medical images, FRD (Fréchet Radiomic Distance), which utilizes standardized, clinically meaningful, and interpretable image features. We show that FRD is superior to other image distribution metrics for a range of medical imaging applications, including out-of-domain (OOD) detection, the evaluation of image-to-image translation (by correlating more with downstream task performance as well as anatomical consistency and realism), and the evaluation of unconditional image generation. Moreover, FRD offers additional benefits such as stability and computational efficiency at low sample sizes, sensitivity to image corruptions and adversarial attacks, feature interpretability, and correlation with radiologist-perceived image quality. Additionally, we address key gaps in the literature by presenting an extensive framework for the multifaceted evaluation of image similarity metrics in medical imaging—including the first large-scale comparative study of generative models for medical image translation—and release an accessible codebase to facilitate future research. Our results are supported by thorough experiments spanning a variety of datasets, modalities, and downstream tasks, highlighting the broad potential of FRD for medical image analysis.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103943"},"PeriodicalIF":11.8,"publicationDate":"2026-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146048368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Medical image analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1