首页 > 最新文献

Medical image analysis最新文献

英文 中文
Physics-informed graph neural networks for flow field estimation in carotid arteries 用于颈动脉流场估计的物理信息图神经网络
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-02-07 DOI: 10.1016/j.media.2026.103974
Julian Suk , Dieuwertje Alblas , Barbara A. Hutten , Albert Wiegman , Christoph Brune , Pim van Ooij , Jelmer M. Wolterink
Hemodynamic quantities are valuable biomedical risk factors for cardiovascular pathology such as atherosclerosis. Non-invasive, in-vivo measurement of these quantities can only be performed using a select number of modalities that are not widely available, such as 4D flow magnetic resonance imaging (MRI). In this work, we create a surrogate model for hemodynamic flow field estimation, powered by machine learning. We train graph neural networks that include priors about the underlying symmetries and physics, limiting the amount of data required for training. This allows us to train the model using moderately-sized, in-vivo 4D flow MRI datasets, instead of large in-silico datasets obtained by computational fluid dynamics (CFD), as is the current standard. We create an efficient, equivariant neural network by combining the popular PointNet++ architecture with group-steerable layers. To incorporate the physics-informed priors, we derive an efficient discretisation scheme for the involved differential operators. We perform extensive experiments in carotid arteries and show that our model can accurately estimate low-noise hemodynamic flow fields in the carotid artery. Moreover, we show how the learned relation between geometry and hemodynamic quantities transfers to 3D vascular models obtained using a different imaging modality than the training data. This shows that physics-informed graph neural networks can be trained using 4D flow MRI data to estimate blood flow in unseen carotid artery geometries.
血液动力学是心血管疾病如动脉粥样硬化的重要生物医学危险因素。这些量的非侵入性体内测量只能使用一些不广泛使用的模式进行,例如4D流磁共振成像(MRI)。在这项工作中,我们创建了一个由机器学习驱动的血流动力学流场估计代理模型。我们训练包含关于底层对称性和物理先验的图神经网络,限制了训练所需的数据量。这使我们能够使用中等大小的体内4D流MRI数据集来训练模型,而不是使用当前标准的计算流体动力学(CFD)获得的大型计算机数据集。我们通过结合流行的PointNet++架构和组导向层,创建了一个高效的等变神经网络。为了结合物理信息先验,我们为所涉及的微分算子推导了一种有效的离散化方案。我们在颈动脉中进行了大量的实验,并表明我们的模型可以准确地估计颈动脉中的低噪声血流动力学流场。此外,我们展示了几何和血流动力学量之间的学习关系如何转移到使用不同的成像方式而不是训练数据获得的3D血管模型。这表明,物理信息图神经网络可以使用4D流MRI数据进行训练,以估计未见的颈动脉几何形状的血流量。
{"title":"Physics-informed graph neural networks for flow field estimation in carotid arteries","authors":"Julian Suk ,&nbsp;Dieuwertje Alblas ,&nbsp;Barbara A. Hutten ,&nbsp;Albert Wiegman ,&nbsp;Christoph Brune ,&nbsp;Pim van Ooij ,&nbsp;Jelmer M. Wolterink","doi":"10.1016/j.media.2026.103974","DOIUrl":"10.1016/j.media.2026.103974","url":null,"abstract":"<div><div>Hemodynamic quantities are valuable biomedical risk factors for cardiovascular pathology such as atherosclerosis. Non-invasive, in-vivo measurement of these quantities can only be performed using a select number of modalities that are not widely available, such as 4D flow magnetic resonance imaging (MRI). In this work, we create a surrogate model for hemodynamic flow field estimation, powered by machine learning. We train graph neural networks that include priors about the underlying symmetries and physics, limiting the amount of data required for training. This allows us to train the model using moderately-sized, in-vivo 4D flow MRI datasets, instead of large in-silico datasets obtained by computational fluid dynamics (CFD), as is the current standard. We create an efficient, equivariant neural network by combining the popular PointNet++ architecture with group-steerable layers. To incorporate the physics-informed priors, we derive an efficient discretisation scheme for the involved differential operators. We perform extensive experiments in carotid arteries and show that our model can accurately estimate low-noise hemodynamic flow fields in the carotid artery. Moreover, we show how the learned relation between geometry and hemodynamic quantities transfers to 3D vascular models obtained using a different imaging modality than the training data. This shows that physics-informed graph neural networks can be trained using 4D flow MRI data to estimate blood flow in unseen carotid artery geometries.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103974"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146138282","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Channel-wise joint disentanglement representation learning for B-mode and super-resolution ultrasound based CAD of breast cancer 基于b模式和超分辨率超声的乳腺癌CAD的通道联合解缠表示学习
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-01-22 DOI: 10.1016/j.media.2026.103957
Yuhang Zheng , Jiale Xu , Qing Hua , Xiaohong Jia , Xueqin Hou , Yanfeng Yao , Zheng Wei , Yulu Zhang , Fanggang Wu , Wei Guo , Yuan Tian , Jun Wang , Shujun Xia , Yijie Dong , Jun Shi , Jianqiao Zhou
B-mode ultrasound (BUS) is widely used in breast cancer diagnosis, while the emerging super-resolution ultrasound (SRUS) provides microvascular information with high spatial resolution, which has shown great potential in improving breast cancer diagnosis. However, as a new ultrasound modality, its diagnosis remains highly dependent on the clinical experience of sonologists, highlighting the need for reliable computer-aided diagnosis (CAD) approaches. In this work, a novel dual-branch network with a Channel-Wise Joint Disentanglement Representation Learning (CW-JDRL) method is proposed for the multimodal ultrasound-based CAD of breast cancer, where one branch processes BUS and the other analyzes multimodal SRUS data. The CW-JDRL is implemented on the SRUS branch by grouping the final-layer network channels to capture both common and specific properties. It consists of two modules, namely Gradient-guided Disentanglement (GD) module and Gramian-based Contrastive Learning Disentanglement (GCLD) module. The former disentangles with gradient guidance to encourage consistency among common channels and distinctiveness among specific ones, and the latter disentangles common and specific representations by integrating them into a unified contrastive objective. Extensive experiments on a multicenter SRUS dataset demonstrate that the proposed dual-branch network with CW-JDRL achieves superior performance over the compared algorithms and maintains robust generalizability to external data. It suggests not only the effectiveness of SRUS for diagnosis of breast cancer, but also the potential of the proposed CAD model in clinical practice. The codes are publicly available at https://github.com/Zyh-AIUltra/CW-JDRL#
b超(BUS)在乳腺癌诊断中应用广泛,而新兴的超分辨率超声(SRUS)提供了高空间分辨率的微血管信息,在提高乳腺癌诊断方面显示出巨大的潜力。然而,作为一种新的超声模式,其诊断仍然高度依赖于超声医师的临床经验,强调需要可靠的计算机辅助诊断(CAD)方法。在这项工作中,提出了一种新的双分支网络,采用通道智能联合解纠缠表示学习(CW-JDRL)方法,用于基于多模态超声的乳腺癌CAD,其中一个分支处理BUS,另一个分支分析多模态SRUS数据。CW-JDRL是在SRUS分支上通过对最后一层网络通道进行分组来实现的,以捕获公共和特定的属性。它包括两个模块,即梯度引导解纠缠(GD)模块和基于gramian的对比学习解纠缠(GCLD)模块。前者通过梯度引导进行消解,鼓励共同通道之间的一致性和特定通道之间的独特性;后者通过将共同表征和特定表征整合成一个统一的对比目标来消解共同表征和特定表征。在多中心SRUS数据集上的大量实验表明,基于CW-JDRL的双分支网络比所比较的算法具有更好的性能,并保持了对外部数据的鲁棒泛化性。这不仅表明了SRUS对乳腺癌诊断的有效性,而且表明了所提出的CAD模型在临床实践中的潜力。这些代码可在https://github.com/Zyh-AIUltra/CW-JDRL#上公开获取
{"title":"Channel-wise joint disentanglement representation learning for B-mode and super-resolution ultrasound based CAD of breast cancer","authors":"Yuhang Zheng ,&nbsp;Jiale Xu ,&nbsp;Qing Hua ,&nbsp;Xiaohong Jia ,&nbsp;Xueqin Hou ,&nbsp;Yanfeng Yao ,&nbsp;Zheng Wei ,&nbsp;Yulu Zhang ,&nbsp;Fanggang Wu ,&nbsp;Wei Guo ,&nbsp;Yuan Tian ,&nbsp;Jun Wang ,&nbsp;Shujun Xia ,&nbsp;Yijie Dong ,&nbsp;Jun Shi ,&nbsp;Jianqiao Zhou","doi":"10.1016/j.media.2026.103957","DOIUrl":"10.1016/j.media.2026.103957","url":null,"abstract":"<div><div>B-mode ultrasound (BUS) is widely used in breast cancer diagnosis, while the emerging super-resolution ultrasound (SRUS) provides microvascular information with high spatial resolution, which has shown great potential in improving breast cancer diagnosis. However, as a new ultrasound modality, its diagnosis remains highly dependent on the clinical experience of sonologists, highlighting the need for reliable computer-aided diagnosis (CAD) approaches. In this work, a novel dual-branch network with a <strong>C</strong>hannel-<strong>W</strong>ise <strong>J</strong>oint <strong>D</strong>isentanglement <strong>R</strong>epresentation <strong>L</strong>earning (CW-JDRL) method is proposed for the multimodal ultrasound-based CAD of breast cancer, where one branch processes BUS and the other analyzes multimodal SRUS data. The CW-JDRL is implemented on the SRUS branch by grouping the final-layer network channels to capture both common and specific properties. It consists of two modules, namely Gradient-guided Disentanglement (GD) module and Gramian-based Contrastive Learning Disentanglement (GCLD) module. The former disentangles with gradient guidance to encourage consistency among common channels and distinctiveness among specific ones, and the latter disentangles common and specific representations by integrating them into a unified contrastive objective. Extensive experiments on a multicenter SRUS dataset demonstrate that the proposed dual-branch network with CW-JDRL achieves superior performance over the compared algorithms and maintains robust generalizability to external data. It suggests not only the effectiveness of SRUS for diagnosis of breast cancer, but also the potential of the proposed CAD model in clinical practice. The codes are publicly available at <span><span>https://github.com/Zyh-AIUltra/CW-JDRL#</span><svg><path></path></svg></span></div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103957"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146032815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
4D monocular surgical reconstruction under arbitrary camera motions 任意摄像机运动下的四维单眼手术重建
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-02-11 DOI: 10.1016/j.media.2026.103989
Jiwei Shan , Zeyu Cai , Cheng-Tai Hsieh , Yirui Li , Hao Liu , Lijun Han , Hesheng Wang , Shing Shin Cheng
Reconstructing deformable surgical scenes from endoscopic videos is a challenging task with important clinical applications. Recent state-of-the-art approaches, such as those based on implicit neural representations or 3D Gaussian splatting, have made notable progress in this area. However, most existing methods are designed for deformable scenes with fixed endoscope viewpoints and rely on stereo depth priors or accurate structure-from-motion for both initialization and optimization. This limits their ability to handle monocular sequences with large camera movements, restricting their use in real clinical settings. To address these limitations, we propose Local-EndoGS, a high-quality 4D reconstruction framework for monocular endoscopic sequences with arbitrary camera motion. Local-EndoGS introduces a progressive, window-based global scene representation that allocates local deformable scene representations for each observed window, enabling scalability to long sequences with substantial camera movement. To overcome unreliable initialization due to the lack of stereo depth or accurate structure-from-motion, we propose a coarse-to-fine initialization strategy that integrates multi-view geometry, cross-window information, and monocular depth priors, providing a robust foundation for subsequent optimization. In addition, we incorporate long-range 2D pixel trajectory constraints and physical motion priors to improve the physical plausibility of the recovered deformations. We comprehensively evaluate Local-EndoGS on three public endoscopic datasets with deformable scenes and varying camera motions. Local-EndoGS achieves superior performance in both appearance quality and geometry, consistently outperforming state-of-the-art methods. Extensive ablation studies further validate the effectiveness of our key designs. Our code will be released upon acceptance at https://github.com/IRMVLab/Local-EndoGS.
从内窥镜视频中重建可变形的手术场景是一项具有重要临床应用的挑战性任务。最近最先进的方法,如基于隐式神经表示或三维高斯飞溅的方法,在这一领域取得了显著进展。然而,大多数现有方法都是针对具有固定内窥镜视点的可变形场景而设计的,并且依赖于立体深度先验或精确的运动结构来进行初始化和优化。这限制了他们处理大型摄像机运动的单目序列的能力,限制了他们在实际临床环境中的使用。为了解决这些限制,我们提出了Local-EndoGS,一个高质量的四维重建框架,用于任意相机运动的单眼内窥镜序列。local - endogs引入了一种渐进式的、基于窗口的全局场景表示,它为每个观察到的窗口分配局部可变形的场景表示,从而可以扩展到具有大量摄像机移动的长序列。为了克服由于缺乏立体深度或精确的运动结构而导致的不可靠初始化,我们提出了一种从粗到精的初始化策略,该策略集成了多视图几何、交叉窗口信息和单目深度先验,为后续优化提供了坚实的基础。此外,我们还结合了远程二维像素轨迹约束和物理运动先验,以提高恢复变形的物理合理性。我们在三个公共内窥镜数据集上全面评估了Local-EndoGS,这些数据集具有可变形的场景和不同的相机运动。Local-EndoGS在外观质量和几何形状方面都取得了卓越的表现,始终优于最先进的方法。广泛的消融研究进一步验证了我们关键设计的有效性。我们的代码将在接受后在https://github.com/IRMVLab/Local-EndoGS上发布。
{"title":"4D monocular surgical reconstruction under arbitrary camera motions","authors":"Jiwei Shan ,&nbsp;Zeyu Cai ,&nbsp;Cheng-Tai Hsieh ,&nbsp;Yirui Li ,&nbsp;Hao Liu ,&nbsp;Lijun Han ,&nbsp;Hesheng Wang ,&nbsp;Shing Shin Cheng","doi":"10.1016/j.media.2026.103989","DOIUrl":"10.1016/j.media.2026.103989","url":null,"abstract":"<div><div>Reconstructing deformable surgical scenes from endoscopic videos is a challenging task with important clinical applications. Recent state-of-the-art approaches, such as those based on implicit neural representations or 3D Gaussian splatting, have made notable progress in this area. However, most existing methods are designed for deformable scenes with fixed endoscope viewpoints and rely on stereo depth priors or accurate structure-from-motion for both initialization and optimization. This limits their ability to handle monocular sequences with large camera movements, restricting their use in real clinical settings. To address these limitations, we propose Local-EndoGS, a high-quality 4D reconstruction framework for monocular endoscopic sequences with arbitrary camera motion. Local-EndoGS introduces a progressive, window-based global scene representation that allocates local deformable scene representations for each observed window, enabling scalability to long sequences with substantial camera movement. To overcome unreliable initialization due to the lack of stereo depth or accurate structure-from-motion, we propose a coarse-to-fine initialization strategy that integrates multi-view geometry, cross-window information, and monocular depth priors, providing a robust foundation for subsequent optimization. In addition, we incorporate long-range 2D pixel trajectory constraints and physical motion priors to improve the physical plausibility of the recovered deformations. We comprehensively evaluate Local-EndoGS on three public endoscopic datasets with deformable scenes and varying camera motions. Local-EndoGS achieves superior performance in both appearance quality and geometry, consistently outperforming state-of-the-art methods. Extensive ablation studies further validate the effectiveness of our key designs. Our code will be released upon acceptance at <span><span>https://github.com/IRMVLab/Local-EndoGS</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103989"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146160397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A navigation-guided 3D breast ultrasound scanning and reconstruction system for automated multi-lesion spatial localization and diagnosis 一种用于多病灶自动定位与诊断的导航引导三维乳腺超声扫描与重建系统
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-01-28 DOI: 10.1016/j.media.2026.103965
Yi Zhang , Yulin Yan , Kun Wang , Muyu Cai , Yifei Xiang , Yan Guo , Puxun Tu , Tao Ying , Xiaojun Chen
Handheld ultrasound (HHUS) is indispensable for breast cancer screening but remains compromised by operator-dependent acquisition, subjective 2D interpretation and clock-face annotation. Existing spatial tracking systems for HHUS typically lack integration, adaptability, flexibility, and robust 3D representation. Additionally, current deep learning diagnostic methods are predominantly based on single ultrasound images, whereas video-based malignancy classification approaches suffer from limited temporal interpretability. In this study, we develop an intelligent navigation-guided breast ultrasound scanning system delivering seamless 3D reconstruction, nipple-centric lesion localization, and video-based malignancy prediction with full adaptation to the routine workflow. Specifically, a Hybrid Lesion-informed Spatiotemporal Transformer (HLST) is proposed to selectively fuse intra- and peri-lesional dynamics augmented from a prompt-driven BUS-SAM-2 foundation model for sequence-level classification. Moreover, a geometry-adaptive clock projection and analysis method is designed to enable automated standardized clock-face orientation and lesion-to-nipple distance measurement for breasts of arbitrary shape, eliminating patient-attached fiducials or pre-marked landmarks. Validation on three breast phantoms demonstrated high correlations with CT reference (r > 0.99 for distance, r > 0.97 for 3D size, and r=1.00 for clockwise angle, p < 0.0001). Clinical evaluation in 43 female patients (30 abnormal breasts) yielded median clock-face orientation and size discrepancies of 0 h and 0.7 mm  ×  0.6 mm, respectively, versus conventional reports. Meanwhile, HLST achieved superior performance (86.1% accuracy) on the BUV dataset. By coupling precise 3D spatial annotation with foundation-model-enhanced spatiotemporal characterization, the proposed system offers a reliable, streamlined workflow that standardizes follow-up, guides biopsies, and promotes diagnostic confidence in HHUS practice.
手持式超声(HHUS)在乳腺癌筛查中是不可或缺的,但仍然受到操作者依赖的获取、主观2D解释和时钟面注释的影响。现有的hus空间跟踪系统通常缺乏集成、适应性、灵活性和鲁棒的3D表示。此外,目前的深度学习诊断方法主要基于单个超声图像,而基于视频的恶性肿瘤分类方法在时间上的可解释性有限。在这项研究中,我们开发了一种智能导航引导的乳房超声扫描系统,提供无缝的3D重建,乳头中心病变定位和基于视频的恶性肿瘤预测,完全适应常规工作流程。具体而言,提出了一种混合病变信息时空转换器(HLST),以选择性地融合从提示驱动的BUS-SAM-2基础模型增强的病变内和病变周围动态,用于序列级分类。此外,设计了一种几何自适应时钟投影和分析方法,可以实现任意形状乳房的自动标准化时钟面方向和病变到乳头距离测量,消除了患者附加的基准或预先标记的地标。三个乳房幻影的验证显示与CT参考高度相关(距离r >; 0.99,3D尺寸r >; 0.97,顺时针角度r=1.00, p <; 0.0001)。43例女性患者(30例异常乳房)的临床评估结果显示,与传统报告相比,时钟面中位取向和尺寸差异分别为0小时和0.7 mm × 0.6 mm。同时,HLST在BUV数据集上取得了86.1%的准确率。通过将精确的3D空间注释与基础模型增强的时空特征相结合,该系统提供了一个可靠的、简化的工作流程,可以标准化随访,指导活检,并提高HHUS实践中的诊断信心。
{"title":"A navigation-guided 3D breast ultrasound scanning and reconstruction system for automated multi-lesion spatial localization and diagnosis","authors":"Yi Zhang ,&nbsp;Yulin Yan ,&nbsp;Kun Wang ,&nbsp;Muyu Cai ,&nbsp;Yifei Xiang ,&nbsp;Yan Guo ,&nbsp;Puxun Tu ,&nbsp;Tao Ying ,&nbsp;Xiaojun Chen","doi":"10.1016/j.media.2026.103965","DOIUrl":"10.1016/j.media.2026.103965","url":null,"abstract":"<div><div>Handheld ultrasound (HHUS) is indispensable for breast cancer screening but remains compromised by operator-dependent acquisition, subjective 2D interpretation and clock-face annotation. Existing spatial tracking systems for HHUS typically lack integration, adaptability, flexibility, and robust 3D representation. Additionally, current deep learning diagnostic methods are predominantly based on single ultrasound images, whereas video-based malignancy classification approaches suffer from limited temporal interpretability. In this study, we develop an intelligent navigation-guided breast ultrasound scanning system delivering seamless 3D reconstruction, nipple-centric lesion localization, and video-based malignancy prediction with full adaptation to the routine workflow. Specifically, a Hybrid Lesion-informed Spatiotemporal Transformer (HLST) is proposed to selectively fuse intra- and peri-lesional dynamics augmented from a prompt-driven BUS-SAM-2 foundation model for sequence-level classification. Moreover, a geometry-adaptive clock projection and analysis method is designed to enable automated standardized clock-face orientation and lesion-to-nipple distance measurement for breasts of arbitrary shape, eliminating patient-attached fiducials or pre-marked landmarks. Validation on three breast phantoms demonstrated high correlations with CT reference (<em>r</em> &gt; 0.99 for distance, <em>r</em> &gt; 0.97 for 3D size, and <span><math><mrow><mi>r</mi><mo>=</mo><mn>1.00</mn></mrow></math></span> for clockwise angle, <em>p</em> &lt; 0.0001). Clinical evaluation in 43 female patients (30 abnormal breasts) yielded median clock-face orientation and size discrepancies of 0 h and 0.7 mm  ×  0.6 mm, respectively, versus conventional reports. Meanwhile, HLST achieved superior performance (86.1% accuracy) on the BUV dataset. By coupling precise 3D spatial annotation with foundation-model-enhanced spatiotemporal characterization, the proposed system offers a reliable, streamlined workflow that standardizes follow-up, guides biopsies, and promotes diagnostic confidence in HHUS practice.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103965"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146072194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DPFR: Semi-supervised gland segmentation via density perturbation and feature recalibration DPFR:通过密度扰动和特征重新校准的半监督腺体分割
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-01-27 DOI: 10.1016/j.media.2026.103962
Jiejiang Yu, Yu Liu
In recent years, semi-supervised methods have attracted considerable attention in gland segmentation of histopathological images, as they can substantially reduce the annotation data burden for pathologists. The most widely adopted approach is the Mean-Teacher framework based on consistency regularization, which exploits unlabeled data information through consistency regularization constraints. However, due to the morphological complexity of glands in histopathological images, existing methods still suffer from confusion between glands and background, as well as gland adhesion. To address these challenges, we propose a semi-supervised gland segmentation method based on Density Perturbation and Feature Recalibration (DPFR). Specifically, we first design a normalized flow-based density estimator to effectively model the feature density distributions of glands, contours, and background. The gradient information of the estimator is then exploited to determine the descent direction in low-density regions, along which perturbations are applied to enhance feature discriminability. Furthermore, a contrastive-learning-based feature recalibration module is designed to alleviate inter-class distribution confusion, thereby improving gland-background separability and mitigating gland adhesion. Extensive experiments on three public gland segmentation datasets demonstrate that the proposed method consistently outperforms existing semi-supervised approaches, achieving state-of-the-art performance with a substantial margin. The code repository address is https://github.com/Methow0/DPFR.
近年来,半监督方法在组织病理图像的腺体分割中引起了广泛的关注,因为它可以大大减少病理学家的注释数据负担。采用最广泛的方法是基于一致性正则化的Mean-Teacher框架,该框架通过一致性正则化约束来利用未标记的数据信息。然而,由于组织病理图像中腺体形态的复杂性,现有方法存在腺体与背景混淆、腺体粘连等问题。为了解决这些问题,我们提出了一种基于密度扰动和特征再校准(DPFR)的半监督腺体分割方法。具体来说,我们首先设计了一个归一化的基于流的密度估计器,以有效地模拟腺体、轮廓和背景的特征密度分布。然后利用估计器的梯度信息确定低密度区域的下降方向,沿此方向施加扰动以增强特征的可判别性。此外,设计了基于对比学习的特征再校准模块,以缓解类间分布混乱,从而提高腺体-背景可分离性,减轻腺体粘附。在三个公共腺体分割数据集上进行的大量实验表明,所提出的方法始终优于现有的半监督方法,在很大程度上实现了最先进的性能。代码存储库地址是https://github.com/Methow0/DPFR。
{"title":"DPFR: Semi-supervised gland segmentation via density perturbation and feature recalibration","authors":"Jiejiang Yu,&nbsp;Yu Liu","doi":"10.1016/j.media.2026.103962","DOIUrl":"10.1016/j.media.2026.103962","url":null,"abstract":"<div><div>In recent years, semi-supervised methods have attracted considerable attention in gland segmentation of histopathological images, as they can substantially reduce the annotation data burden for pathologists. The most widely adopted approach is the Mean-Teacher framework based on consistency regularization, which exploits unlabeled data information through consistency regularization constraints. However, due to the morphological complexity of glands in histopathological images, existing methods still suffer from confusion between glands and background, as well as gland adhesion. To address these challenges, we propose a semi-supervised gland segmentation method based on Density Perturbation and Feature Recalibration (DPFR). Specifically, we first design a normalized flow-based density estimator to effectively model the feature density distributions of glands, contours, and background. The gradient information of the estimator is then exploited to determine the descent direction in low-density regions, along which perturbations are applied to enhance feature discriminability. Furthermore, a contrastive-learning-based feature recalibration module is designed to alleviate inter-class distribution confusion, thereby improving gland-background separability and mitigating gland adhesion. Extensive experiments on three public gland segmentation datasets demonstrate that the proposed method consistently outperforms existing semi-supervised approaches, achieving state-of-the-art performance with a substantial margin. The code repository address is <span><span>https://github.com/Methow0/DPFR</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103962"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146072196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incorporating global-local tissue changes to predict future breast cancer from longitudinal screening mammograms 结合整体-局部组织变化,通过乳房x线纵向筛查预测未来乳腺癌
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-02-16 DOI: 10.1016/j.media.2026.103990
Xin Wang , Tao Tan , Yuan Gao , Eric Marcus , Hong-Yu Zhou , Chunyao Lu , Luyi Han , Antonio Portaluri , Ruisheng Su , Tianyu Zhang , Xinglong Liang , Regina Beets-Tan , Katja Pinker , Yue Sun , Ritse Mann , Jonas Teuwen
Early detection of breast cancer (BC) through mammography screening is critical for reducing mortality and improving patient outcomes. However, full-population-based, age-driven screening might not lead to optimal resource use and may enlarge screening associated harms in low risk women. Accurate and interpretable BC risk prediction is essential to improve strategies and make screening more personalized. Although recent deep learning models have shown promise in leveraging mammograms for risk stratification, challenges remain in interpretable modeling of temporal changes, efficiently capturing multi-scale risk tissue features from large-scale images, and precise time prediction to enhance clinical interpretability. In this study, we propose Tarcking-Aware Breast Cancer Risk model (TA-BreaCR), a novel framework that integrates local-to-global multiscale longitudinal tissue changes and explicitly models the ordinal relationship of time to BC events, enabling joint prediction of both future BC risk and estimated time to onset. The model is evaluated on two datasets (In-house and EMBED), outperforming existing and state-of-the-art methods in both risk classification and time-to-event prediction tasks. Visualization analysis reveals consistent attention to high-risk regions over time, enhancing interpretability. These results highlight the potential of TA-BreaCR to support individualized BC screening and prevention.
通过乳房x光检查早期发现乳腺癌(BC)对于降低死亡率和改善患者预后至关重要。然而,完全以人群为基础、年龄驱动的筛查可能不会导致资源的最佳利用,并可能扩大筛查对低风险妇女的相关危害。准确和可解释的BC风险预测对于改进策略和使筛查更加个性化至关重要。尽管最近的深度学习模型在利用乳房x线照片进行风险分层方面显示出了希望,但在时间变化的可解释建模、从大规模图像中有效捕获多尺度风险组织特征以及精确的时间预测以提高临床可解释性方面仍然存在挑战。在这项研究中,我们提出了追踪感知乳腺癌风险模型(TA-BreaCR),这是一个新颖的框架,整合了局部到全球的多尺度纵向组织变化,明确地模拟了时间与BC事件的顺序关系,从而能够联合预测未来的BC风险和估计的发病时间。该模型在两个数据集(内部和嵌入)上进行了评估,在风险分类和事件时间预测任务方面优于现有和最先进的方法。可视化分析揭示了对高风险区域的持续关注,增强了可解释性。这些结果突出了TA-BreaCR在支持个体化BC筛查和预防方面的潜力。
{"title":"Incorporating global-local tissue changes to predict future breast cancer from longitudinal screening mammograms","authors":"Xin Wang ,&nbsp;Tao Tan ,&nbsp;Yuan Gao ,&nbsp;Eric Marcus ,&nbsp;Hong-Yu Zhou ,&nbsp;Chunyao Lu ,&nbsp;Luyi Han ,&nbsp;Antonio Portaluri ,&nbsp;Ruisheng Su ,&nbsp;Tianyu Zhang ,&nbsp;Xinglong Liang ,&nbsp;Regina Beets-Tan ,&nbsp;Katja Pinker ,&nbsp;Yue Sun ,&nbsp;Ritse Mann ,&nbsp;Jonas Teuwen","doi":"10.1016/j.media.2026.103990","DOIUrl":"10.1016/j.media.2026.103990","url":null,"abstract":"<div><div>Early detection of breast cancer (BC) through mammography screening is critical for reducing mortality and improving patient outcomes. However, full-population-based, age-driven screening might not lead to optimal resource use and may enlarge screening associated harms in low risk women. Accurate and interpretable BC risk prediction is essential to improve strategies and make screening more personalized. Although recent deep learning models have shown promise in leveraging mammograms for risk stratification, challenges remain in interpretable modeling of temporal changes, efficiently capturing multi-scale risk tissue features from large-scale images, and precise time prediction to enhance clinical interpretability. In this study, we propose <u>T</u>arcking-<u>A</u>ware <u>Brea</u>st <u>C</u>ancer <u>R</u>isk model (TA-BreaCR), a novel framework that integrates local-to-global multiscale longitudinal tissue changes and explicitly models the ordinal relationship of time to BC events, enabling joint prediction of both future BC risk and estimated time to onset. The model is evaluated on two datasets (In-house and EMBED), outperforming existing and state-of-the-art methods in both risk classification and time-to-event prediction tasks. Visualization analysis reveals consistent attention to high-risk regions over time, enhancing interpretability. These results highlight the potential of TA-BreaCR to support individualized BC screening and prevention.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103990"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146209647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CATERPillar: a flexible framework for generating white matter numerical substrates with incorporated glial cells 卡特彼勒:一个灵活的框架,生成白质数值基质与合并胶质细胞
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-01-17 DOI: 10.1016/j.media.2026.103946
Jasmine Nguyen-Duc , Malte Brammerloh , Melina Cherchali , Inès De Riedmatten , Jean-Baptiste Pérot , Jonathan Rafael-Patiño , Ileana O. Jelescu
Monte Carlo diffusion simulations in numerical substrates are valuable for exploring the sensitivity and specificity of the diffusion MRI (dMRI) signal to realistic cell microstructure features. A crucial component of such simulations is the use of numerical phantoms that accurately represent the target tissue, which is in this case, cerebral white matter (WM). This study introduces CATERPillar (Computational Axonal Threading Engine for Realistic Proliferation), a novel method that simulates the mechanics of axonal growth using overlapping spheres as elementary units. CATERPillar facilitates parallel axon development while preventing collisions, offering user control over key structural parameters such as cellular density, undulation, beading and myelination. Its uniqueness lies in its ability to generate not only realistic axonal structures but also realistic glial cells, enhancing the biological fidelity of simulations. We showed that our grown substrates feature distributions of key morphological parameters that agree with those from histological studies. The structural realism of the astrocytic components was quantitatively validated using Sholl analysis. Furthermore, the time-dependent diffusion in the extra- and intra-axonal compartments accurately reflected expected characteristics of short-range disorder, as predicted by theoretical models. CATERPillar is open source and can be used to (a) develop new acquisition schemes that sensitise the MRI signal to unique tissue microstructure features, (b) test the accuracy of a broad range of analytical models, and (c) build a set of substrates to train machine learning models on.
在数值基底上进行蒙特卡罗扩散模拟对于探索扩散核磁共振成像(dMRI)信号对真实细胞微观结构特征的敏感性和特异性具有重要意义。这种模拟的一个关键组成部分是使用数字幻象来准确地代表目标组织,在这种情况下,就是脑白质(WM)。本文介绍了一种以重叠球体为基本单元模拟轴突生长机制的新方法——毛毛虫(Computational Axonal Threading Engine for Realistic Proliferation)。卡特彼勒促进平行轴突的发育,同时防止碰撞,为用户提供对关键结构参数的控制,如细胞密度、波动、串珠和髓鞘形成。它的独特之处在于它不仅能够生成真实的轴突结构,而且能够生成真实的胶质细胞,从而提高了模拟的生物保真度。我们发现,我们培养的底物具有与组织学研究一致的关键形态参数分布。星形细胞成分的结构真实性被定量验证使用肖尔分析。此外,轴突外室和轴突内室的时间依赖性扩散准确地反映了理论模型预测的短期紊乱的预期特征。卡特彼勒是开源的,可用于(a)开发新的采集方案,使MRI信号对独特的组织微观结构特征敏感,(b)测试广泛分析模型的准确性,以及(c)构建一组基板来训练机器学习模型。
{"title":"CATERPillar: a flexible framework for generating white matter numerical substrates with incorporated glial cells","authors":"Jasmine Nguyen-Duc ,&nbsp;Malte Brammerloh ,&nbsp;Melina Cherchali ,&nbsp;Inès De Riedmatten ,&nbsp;Jean-Baptiste Pérot ,&nbsp;Jonathan Rafael-Patiño ,&nbsp;Ileana O. Jelescu","doi":"10.1016/j.media.2026.103946","DOIUrl":"10.1016/j.media.2026.103946","url":null,"abstract":"<div><div>Monte Carlo diffusion simulations in numerical substrates are valuable for exploring the sensitivity and specificity of the diffusion MRI (dMRI) signal to realistic cell microstructure features. A crucial component of such simulations is the use of numerical phantoms that accurately represent the target tissue, which is in this case, cerebral white matter (WM). This study introduces CATERPillar (Computational Axonal Threading Engine for Realistic Proliferation), a novel method that simulates the mechanics of axonal growth using overlapping spheres as elementary units. CATERPillar facilitates parallel axon development while preventing collisions, offering user control over key structural parameters such as cellular density, undulation, beading and myelination. Its uniqueness lies in its ability to generate not only realistic axonal structures but also realistic glial cells, enhancing the biological fidelity of simulations. We showed that our grown substrates feature distributions of key morphological parameters that agree with those from histological studies. The structural realism of the astrocytic components was quantitatively validated using Sholl analysis. Furthermore, the time-dependent diffusion in the extra- and intra-axonal compartments accurately reflected expected characteristics of short-range disorder, as predicted by theoretical models. CATERPillar is open source and can be used to (a) develop new acquisition schemes that sensitise the MRI signal to unique tissue microstructure features, (b) test the accuracy of a broad range of analytical models, and (c) build a set of substrates to train machine learning models on.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103946"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145995244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anatomy-guided prompting with cross-modal self-alignment for whole-body PET-CT breast cancer segmentation 解剖引导提示与跨模态自对准用于全身PET-CT乳腺癌分割
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-01-22 DOI: 10.1016/j.media.2026.103956
Jiaju Huang , Xiao Yang , Xinglong Liang , Shaobin Chen , Yue Sun , Greta Sp Mok , Shuo Li , Ying Wang , Tao Tan
Accurate segmentation of breast cancer in PET-CT images is crucial for precise staging, monitoring treatment response, and guiding personalized therapy. However, the small size and dispersed nature of metastatic lesions, coupled with the scarcity of annotated data and heterogeneity between modalities that hinders effective information fusion, make this task challenging. This paper proposes a novel anatomy-guided cross-modal learning framework to address these issues. Our approach first generates organ pseudo-labels through a teacher-student learning paradigm, which serve as anatomical prompts to guide cancer segmentation. We then introduce a self-aligning cross-modal pre-training method that aligns PET and CT features in a shared latent space through masked 3D patch reconstruction, enabling effective cross-modal feature fusion. Finally, we initialize the segmentation network’s encoder with the pre-trained encoder weights, and incorporate organ labels through a Mamba-based prompt encoder and Hypernet-Controlled Cross-Attention mechanism for dynamic anatomical feature extraction and fusion. Notably, our method outperforms eight state-of-the-art methods, including CNN-based, transformer-based, and Mamba-based approaches, on two datasets encompassing primary breast cancer, metastatic breast cancer, and other types of cancer segmentation tasks.
在PET-CT图像中准确分割乳腺癌对于精确分期、监测治疗反应和指导个性化治疗至关重要。然而,转移病灶的小尺寸和分散性质,加上注释数据的缺乏和模式之间的异质性,阻碍了有效的信息融合,使得这项任务具有挑战性。本文提出了一种新的解剖学指导的跨模态学习框架来解决这些问题。我们的方法首先通过师生学习范式生成器官伪标签,作为指导癌症分割的解剖提示。然后,我们引入了一种自对齐的跨模态预训练方法,该方法通过掩膜3D补丁重建在共享潜在空间中对齐PET和CT特征,从而实现有效的跨模态特征融合。最后,我们使用预训练的编码器权重初始化分割网络的编码器,并通过基于mamba的提示编码器和hypernet控制的交叉注意机制合并器官标签,进行动态解剖特征提取和融合。值得注意的是,我们的方法在包括原发性乳腺癌、转移性乳腺癌和其他类型的癌症分割任务的两个数据集上优于八种最先进的方法,包括基于cnn、基于transformer和基于mamba的方法。
{"title":"Anatomy-guided prompting with cross-modal self-alignment for whole-body PET-CT breast cancer segmentation","authors":"Jiaju Huang ,&nbsp;Xiao Yang ,&nbsp;Xinglong Liang ,&nbsp;Shaobin Chen ,&nbsp;Yue Sun ,&nbsp;Greta Sp Mok ,&nbsp;Shuo Li ,&nbsp;Ying Wang ,&nbsp;Tao Tan","doi":"10.1016/j.media.2026.103956","DOIUrl":"10.1016/j.media.2026.103956","url":null,"abstract":"<div><div>Accurate segmentation of breast cancer in PET-CT images is crucial for precise staging, monitoring treatment response, and guiding personalized therapy. However, the small size and dispersed nature of metastatic lesions, coupled with the scarcity of annotated data and heterogeneity between modalities that hinders effective information fusion, make this task challenging. This paper proposes a novel anatomy-guided cross-modal learning framework to address these issues. Our approach first generates organ pseudo-labels through a teacher-student learning paradigm, which serve as anatomical prompts to guide cancer segmentation. We then introduce a self-aligning cross-modal pre-training method that aligns PET and CT features in a shared latent space through masked 3D patch reconstruction, enabling effective cross-modal feature fusion. Finally, we initialize the segmentation network’s encoder with the pre-trained encoder weights, and incorporate organ labels through a Mamba-based prompt encoder and Hypernet-Controlled Cross-Attention mechanism for dynamic anatomical feature extraction and fusion. Notably, our method outperforms eight state-of-the-art methods, including CNN-based, transformer-based, and Mamba-based approaches, on two datasets encompassing primary breast cancer, metastatic breast cancer, and other types of cancer segmentation tasks.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103956"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146033893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Knowledge-guided multi-geometric window transformer for cardiac cine MRI reconstruction 知识引导的心脏MRI重建多几何窗口变压器
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-01-09 DOI: 10.1016/j.media.2026.103936
Jun Lyu , Guangming Wang , Yunqi Wang , Jing Qin , Chengyan Wang
Magnetic resonance imaging (MRI) plays a crucial role in clinical diagnosis, yet traditional MR image acquisition often requires a prolonged duration, potentially causing patient discomfort and image artifacts. Faster and more accurate image reconstruction may alleviate patient discomfort during MRI examinations and enhance diagnostic accuracy and efficiency. In recent years, significant advancements in deep learning technology offer promise for improving MR image quality and accelerating acquisition. Addressing the demand for cardiac cine MRI reconstruction, we propose KGMgT, a novel MRI reconstruction network based on knowledge-guided approaches. The KGMgT model leverages adaptive spatiotemporal attention mechanisms to infer motion trajectories of adjacent cardiac frames, thereby better extracting complementary information. Additionally, we employ Transformer-driven dynamic feature aggregation to establish long-range dependencies, facilitating global information integration. Research findings demonstrate that the KGMgT model achieves state-of-the-art performance on multiple benchmark datasets, offering an efficient solution for cardiac cine MRI reconstruction. This collaborative approach, combining artificial intelligence technology to assist medical professionals in clinical decision-making, holds promise for significantly improving diagnostic efficiency, optimizing treatment plans, and enhancing the patient treatment experience. The code and trained models are available at https://github.com/MICV-Lab/KGMgT.
磁共振成像(MRI)在临床诊断中起着至关重要的作用,然而传统的磁共振图像采集通常需要较长的时间,可能会导致患者不适和图像伪影。更快更准确的图像重建可以减轻患者在MRI检查时的不适,提高诊断的准确性和效率。近年来,深度学习技术的重大进步为提高MR图像质量和加速采集提供了希望。针对心脏电影MRI重构的需求,我们提出了一种基于知识引导方法的新型MRI重构网络KGMgT。KGMgT模型利用自适应时空注意机制来推断相邻心脏帧的运动轨迹,从而更好地提取互补信息。此外,我们使用transformer驱动的动态特征聚合来建立远程依赖,促进全局信息集成。研究结果表明,KGMgT模型在多个基准数据集上达到了最先进的性能,为心脏影像MRI重建提供了有效的解决方案。这种协作方法结合人工智能技术来协助医疗专业人员进行临床决策,有望显著提高诊断效率,优化治疗方案,并增强患者的治疗体验。代码和经过训练的模型可在https://github.com/MICV-Lab/KGMgT上获得。
{"title":"Knowledge-guided multi-geometric window transformer for cardiac cine MRI reconstruction","authors":"Jun Lyu ,&nbsp;Guangming Wang ,&nbsp;Yunqi Wang ,&nbsp;Jing Qin ,&nbsp;Chengyan Wang","doi":"10.1016/j.media.2026.103936","DOIUrl":"10.1016/j.media.2026.103936","url":null,"abstract":"<div><div>Magnetic resonance imaging (MRI) plays a crucial role in clinical diagnosis, yet traditional MR image acquisition often requires a prolonged duration, potentially causing patient discomfort and image artifacts. Faster and more accurate image reconstruction may alleviate patient discomfort during MRI examinations and enhance diagnostic accuracy and efficiency. In recent years, significant advancements in deep learning technology offer promise for improving MR image quality and accelerating acquisition. Addressing the demand for cardiac cine MRI reconstruction, we propose KGMgT, a novel MRI reconstruction network based on knowledge-guided approaches. The KGMgT model leverages adaptive spatiotemporal attention mechanisms to infer motion trajectories of adjacent cardiac frames, thereby better extracting complementary information. Additionally, we employ Transformer-driven dynamic feature aggregation to establish long-range dependencies, facilitating global information integration. Research findings demonstrate that the KGMgT model achieves state-of-the-art performance on multiple benchmark datasets, offering an efficient solution for cardiac cine MRI reconstruction. This collaborative approach, combining artificial intelligence technology to assist medical professionals in clinical decision-making, holds promise for significantly improving diagnostic efficiency, optimizing treatment plans, and enhancing the patient treatment experience. The code and trained models are available at <span><span>https://github.com/MICV-Lab/KGMgT</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103936"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145956958","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards boundary confusion for volumetric medical image segmentation 体医学图像分割中边界混淆的研究
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-01 Epub Date: 2026-01-25 DOI: 10.1016/j.media.2026.103961
Xin You , Ming Ding , Minghui Zhang , Hanxiao Zhang , Junyang Wu , Yi Yu , Jie Yang , Yun Gu
Accurate boundary segmentation of volumetric images is a critical task for image-guided diagnosis and computer-assisted intervention. It is challenging to address the boundary confusion with explicit constraints. Existing methods of refining boundaries overemphasize the slender structure while overlooking the dynamic interactions between boundaries and neighboring regions. In this paper, we reconceptualize the mechanism of boundary generation via introducing Pushing and Pulling interactions, then propose a unified network termed PP-Net to model shape characteristics of the confused boundary region. Specifically, we first propose the semantic difference module (SDM) from the pushing branch to drive the boundary towards the ground truth under diffusion guidance. Additionally, the class clustering module (CCM) from the pulling branch is introduced to stretch the intersected boundary along the opposite direction. Thus, pushing and pulling branches will furnish two adversarial forces to enhance representation capabilities for the faint boundary. Experiments are conducted on four public datasets and one in-house dataset plagued by boundary confusion. The results demonstrate the superiority of PP-Net over other segmentation networks, especially on the evaluation metrics of Hausdorff Distance and Average Symmetric Surface Distance. Besides, SDM and CCM can serve as plug-and-play modules to enhance classic U-shape baseline models, including recent SAM-based foundation models. Source codes are available at https://github.com/EndoluminalSurgicalVision-IMR/PnPNet.
体积图像的精确边界分割是图像引导诊断和计算机辅助干预的关键任务。用明确的约束来解决边界混淆是具有挑战性的。现有的边界细化方法过分强调细长的结构,而忽略了边界与邻近区域之间的动态相互作用。本文通过引入推拉相互作用,重新定义了边界生成的机制,并提出了一个统一的PP-Net网络来模拟混乱边界区域的形状特征。具体而言,我们首先提出了推枝的语义差分模块(SDM),在扩散引导下将边界向地面真值驱动。此外,引入了来自拉分支的类聚类模块(CCM),将相交边界沿相反方向拉伸。因此,推和拉分支将提供两种对立的力量,以增强对模糊边界的表示能力。在四个公共数据集和一个边界混乱的内部数据集上进行了实验。结果表明,PP-Net在Hausdorff距离和平均对称表面距离的评价指标上优于其他分割网络。此外,SDM和CCM可以作为即插即用模块来增强经典的u型基线模型,包括最近基于sam的基础模型。源代码可从https://github.com/EndoluminalSurgicalVision-IMR/PnPNet获得。
{"title":"Towards boundary confusion for volumetric medical image segmentation","authors":"Xin You ,&nbsp;Ming Ding ,&nbsp;Minghui Zhang ,&nbsp;Hanxiao Zhang ,&nbsp;Junyang Wu ,&nbsp;Yi Yu ,&nbsp;Jie Yang ,&nbsp;Yun Gu","doi":"10.1016/j.media.2026.103961","DOIUrl":"10.1016/j.media.2026.103961","url":null,"abstract":"<div><div>Accurate boundary segmentation of volumetric images is a critical task for image-guided diagnosis and computer-assisted intervention. It is challenging to address the boundary confusion with explicit constraints. Existing methods of refining boundaries overemphasize the slender structure while overlooking the dynamic interactions between boundaries and neighboring regions. In this paper, we reconceptualize the mechanism of boundary generation via introducing Pushing and Pulling interactions, then propose a unified network termed PP-Net to model shape characteristics of the confused boundary region. Specifically, we first propose the semantic difference module (SDM) from the pushing branch to drive the boundary towards the ground truth under diffusion guidance. Additionally, the class clustering module (CCM) from the pulling branch is introduced to stretch the intersected boundary along the opposite direction. Thus, pushing and pulling branches will furnish two adversarial forces to enhance representation capabilities for the faint boundary. Experiments are conducted on four public datasets and one in-house dataset plagued by boundary confusion. The results demonstrate the superiority of PP-Net over other segmentation networks, especially on the evaluation metrics of Hausdorff Distance and Average Symmetric Surface Distance. Besides, SDM and CCM can serve as plug-and-play modules to enhance classic U-shape baseline models, including recent SAM-based foundation models. Source codes are available at <span><span>https://github.com/EndoluminalSurgicalVision-IMR/PnPNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103961"},"PeriodicalIF":11.8,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146048365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Medical image analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1