首页 > 最新文献

Medical image analysis最新文献

英文 中文
PISCO: Self-supervised k-space regularization for improved neural implicit k-space representations of dynamic MRI 动态MRI神经隐式k空间表征的自监督k空间正则化
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-29 DOI: 10.1016/j.media.2025.103890
Veronika Spieker , Hannah Eichhorn , Wenqi Huang , Jonathan K. Stelter , Tabita Catalan , Rickmer F. Braren , Daniel Rueckert , Francisco Sahli Costabal , Kerstin Hammernik , Dimitrios C. Karampinos , Claudia Prieto , Julia A. Schnabel
Neural implicit k-space representations (NIK) have shown promising results for dynamic magnetic resonance imaging (MRI) at high temporal resolutions. Yet, reducing acquisition time, and thereby available training data, results in severe performance drops due to overfitting. To address this, we introduce a novel self-supervised k-space loss function LPISCO, applicable for regularization of NIK-based reconstructions. The proposed loss function is based on the concept of parallel imaging-inspired self-consistency (PISCO), enforcing a consistent global k-space neighborhood relationship without requiring additional data. Quantitative and qualitative evaluations on static and dynamic MR reconstructions show that integrating PISCO significantly improves NIK representations, making it a competitive dynamic reconstruction method without constraining the temporal resolution. Particularly at high acceleration factors (R ≥ 50), NIK with PISCO can avoid temporal oversmoothing of state-of-the-art methods and achieves superior spatio-temporal reconstruction quality. Furthermore, an extensive analysis of the loss assumptions and stability shows PISCO’s potential as versatile self-supervised k-space loss function for further applications and architectures. Code is available at: https://github.com/compai-lab/2025-pisco-spieker
神经隐式k空间表示(NIK)在高时间分辨率下的动态磁共振成像(MRI)中显示出有希望的结果。然而,减少获取时间,从而减少可用的训练数据,会导致过度拟合导致严重的性能下降。为了解决这个问题,我们引入了一种新的自监督k空间损失函数LPISCO,适用于基于nik的重构的正则化。所提出的损失函数基于并行成像启发自洽(PISCO)的概念,在不需要额外数据的情况下强制执行一致的全局k空间邻域关系。静态和动态MR重建的定量和定性评价表明,整合PISCO显著改善了NIK表征,使其成为一种具有竞争力的动态重建方法,而不限制时间分辨率。特别是在高加速度因子(R ≥ 50)下,采用PISCO的NIK可以避免现有方法的时间过平滑,获得较好的时空重建质量。此外,对损失假设和稳定性的广泛分析表明,PISCO作为多功能自监督k空间损失函数的潜力可用于进一步的应用和体系结构。代码可从https://github.com/compai-lab/2025-pisco-spieker获得
{"title":"PISCO: Self-supervised k-space regularization for improved neural implicit k-space representations of dynamic MRI","authors":"Veronika Spieker ,&nbsp;Hannah Eichhorn ,&nbsp;Wenqi Huang ,&nbsp;Jonathan K. Stelter ,&nbsp;Tabita Catalan ,&nbsp;Rickmer F. Braren ,&nbsp;Daniel Rueckert ,&nbsp;Francisco Sahli Costabal ,&nbsp;Kerstin Hammernik ,&nbsp;Dimitrios C. Karampinos ,&nbsp;Claudia Prieto ,&nbsp;Julia A. Schnabel","doi":"10.1016/j.media.2025.103890","DOIUrl":"10.1016/j.media.2025.103890","url":null,"abstract":"<div><div>Neural implicit k-space representations (NIK) have shown promising results for dynamic magnetic resonance imaging (MRI) at high temporal resolutions. Yet, reducing acquisition time, and thereby available training data, results in severe performance drops due to overfitting. To address this, we introduce a novel self-supervised k-space loss function <span><math><msub><mi>L</mi><mtext>PISCO</mtext></msub></math></span>, applicable for regularization of NIK-based reconstructions. The proposed loss function is based on the concept of parallel imaging-inspired self-consistency (PISCO), enforcing a consistent global k-space neighborhood relationship without requiring additional data. Quantitative and qualitative evaluations on static and dynamic MR reconstructions show that integrating PISCO significantly improves NIK representations, making it a competitive dynamic reconstruction method without constraining the temporal resolution. Particularly at high acceleration factors (R ≥ 50), NIK with PISCO can avoid temporal oversmoothing of state-of-the-art methods and achieves superior spatio-temporal reconstruction quality. Furthermore, an extensive analysis of the loss assumptions and stability shows PISCO’s potential as versatile self-supervised k-space loss function for further applications and architectures. Code is available at: <span><span>https://github.com/compai-lab/2025-pisco-spieker</span><svg><path></path></svg></span></div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103890"},"PeriodicalIF":11.8,"publicationDate":"2025-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145619566","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GEM-pRF: GPU-empowered mapping of population receptive fields for large-scale fMRI analysis GEM-pRF:基于gpu的大规模功能磁共振成像(fMRI)分析的群体接受域映射
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-28 DOI: 10.1016/j.media.2025.103891
Siddharth Mittal, Michael Woletz, David Linhardt, Christian Windischberger
Population receptive field (pRF) mapping is a fundamental technique for understanding retinotopic organisation of the human visual system. Since its introduction in 2008, however, its scalability has been severely hindered by the computational bottleneck of iterative parameter refinement. Current state-of-the-art implementations either sacrifice precision for speed or rely on slow iterative parameter updates, limiting their applicability to large-scale datasets. Here, we present a novel mathematical reformulation of the General Linear Model (GLM), wrapped in a GPU-Empowered Mapping of population Receptive Fields (GEM-pRF) software implementation. By orthogonalizing the design matrix, our approach enables the direct and fast computation of the objective function’s derivatives, which are used to eliminate the iterative refinement process. This approach dramatically accelerates pRF estimation with high accuracy. Validation using empirical and simulated data confirms GEM-pRF’s accuracy, and benchmarking against established tools demonstrates a reduction in computation time of almost two orders of magnitude. With its modular and extensible design, GEM-pRF provides a critical advancement for large-scale fMRI retinotopic mapping. Furthermore, our reformulated GLM approach in combination with GPU-based implementation offers a broadly applicable solution that may extend beyond visual neuroscience, accelerating computational modelling across various domains in neuroimaging and beyond.
群体感受野(pRF)映射是理解人类视觉系统视网膜组织的基本技术。然而,自2008年推出以来,其可扩展性受到迭代参数细化的计算瓶颈的严重阻碍。当前最先进的实现要么牺牲精度换取速度,要么依赖缓慢的迭代参数更新,限制了它们对大规模数据集的适用性。在这里,我们提出了一个新的通用线性模型(GLM)的数学重新表述,包裹在一个gpu授权的群体接受域映射(GEM-pRF)软件实现中。通过正交化设计矩阵,我们的方法可以直接和快速地计算目标函数的导数,从而消除迭代优化过程。该方法极大地提高了pRF估计的精度。使用经验和模拟数据的验证证实了GEM-pRF的准确性,针对既定工具的基准测试表明,计算时间减少了近两个数量级。由于其模块化和可扩展的设计,GEM-pRF为大规模功能磁共振成像视网膜定位提供了关键的进步。此外,我们重新制定的GLM方法与基于gpu的实现相结合,提供了一种广泛适用的解决方案,可以扩展到视觉神经科学之外,加速神经成像等各个领域的计算建模。
{"title":"GEM-pRF: GPU-empowered mapping of population receptive fields for large-scale fMRI analysis","authors":"Siddharth Mittal,&nbsp;Michael Woletz,&nbsp;David Linhardt,&nbsp;Christian Windischberger","doi":"10.1016/j.media.2025.103891","DOIUrl":"10.1016/j.media.2025.103891","url":null,"abstract":"<div><div>Population receptive field (pRF) mapping is a fundamental technique for understanding retinotopic organisation of the human visual system. Since its introduction in 2008, however, its scalability has been severely hindered by the computational bottleneck of iterative parameter refinement. Current state-of-the-art implementations either sacrifice precision for speed or rely on slow iterative parameter updates, limiting their applicability to large-scale datasets. Here, we present a novel mathematical reformulation of the General Linear Model (GLM), wrapped in a GPU-Empowered Mapping of population Receptive Fields (GEM-pRF) software implementation. By orthogonalizing the design matrix, our approach enables the direct and fast computation of the objective function’s derivatives, which are used to eliminate the iterative refinement process. This approach dramatically accelerates pRF estimation with high accuracy. Validation using empirical and simulated data confirms GEM-pRF’s accuracy, and benchmarking against established tools demonstrates a reduction in computation time of almost two orders of magnitude. With its modular and extensible design, GEM-pRF provides a critical advancement for large-scale fMRI retinotopic mapping. Furthermore, our reformulated GLM approach in combination with GPU-based implementation offers a broadly applicable solution that may extend beyond visual neuroscience, accelerating computational modelling across various domains in neuroimaging and beyond.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103891"},"PeriodicalIF":11.8,"publicationDate":"2025-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145611717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatial transcriptomics expression prediction from histopathology based on cross-modal mask reconstruction and contrastive learning 基于交叉模态掩模重建和对比学习的组织病理学空间转录组学表达预测
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-25 DOI: 10.1016/j.media.2025.103889
Junzhuo Liu , Markus Eckstein , Zhixiang Wang , Friedrich Feuerhake , Dorit Merhof
Spatial transcriptomics is a technology that captures gene expression at different spatial locations, widely used in tumor microenvironment analysis and molecular profiling of histopathology, providing valuable insights into resolving gene expression and clinical diagnosis of cancer. Due to the high cost of data acquisition, large-scale spatial transcriptomics data remain challenging to obtain. In this study, we develop a contrastive learning-based deep learning method to predict spatially resolved gene expression from the whole-slide images (WSIs). Unlike existing end-to-end prediction frameworks, our method leverages multi-modal contrastive learning to establish a correspondence between histopathological morphology and spatial gene expression in the feature space. By computing cross-modal feature similarity, our method generates spatially resolved gene expression directly from WSIs. Furthermore, to enhance the standard contrastive learning paradigm, a cross-modal masked reconstruction is designed as a pretext task, enabling feature-level fusion between modalities. Notably, our method does not rely on large-scale pretraining datasets or abstract semantic representations from either modality, making it particularly effective for scenarios with limited spatial transcriptomics data. Evaluation across six different disease datasets demonstrates that, compared to existing studies, our method improves Pearson Correlation Coefficient (PCC) in the prediction of highly expressed genes, highly variable genes, and marker genes by 6.27 %, 6.11 %, and 11.26 % respectively. Further analysis indicates that our method preserves gene-gene correlations and applies to datasets with limited samples. Additionally, our method exhibits potential in cancer tissue localization based on biomarker expression. The code repository for this work is available at https://github.com/ngfufdrdh/CMRCNet.
空间转录组学是一种捕获不同空间位置基因表达的技术,广泛应用于肿瘤微环境分析和组织病理学分子谱分析,为解决基因表达和癌症临床诊断提供有价值的见解。由于数据采集成本高,获得大规模空间转录组学数据仍然具有挑战性。在这项研究中,我们开发了一种基于对比学习的深度学习方法,从全幻灯片图像(wsi)中预测空间分辨基因表达。与现有的端到端预测框架不同,我们的方法利用多模态对比学习在特征空间中建立组织病理形态和空间基因表达之间的对应关系。通过计算跨模态特征相似度,我们的方法直接从wsi中生成空间分解的基因表达。此外,为了增强标准的对比学习范式,设计了一个跨模态掩模重构作为借口任务,实现了模态之间的特征级融合。值得注意的是,我们的方法不依赖于大规模的预训练数据集,也不依赖于任何一种模式的抽象语义表示,这使得它在空间转录组学数据有限的情况下特别有效。对6个不同疾病数据集的评估表明,与现有研究相比,我们的方法在预测高表达基因、高可变基因和标记基因方面的Pearson相关系数(PCC)分别提高了6.27%、6.11%和11.26%。进一步的分析表明,我们的方法保留了基因-基因的相关性,并适用于有限样本的数据集。此外,我们的方法显示出基于生物标志物表达的癌症组织定位的潜力。这项工作的代码存储库可从https://github.com/ngfufdrdh/CMRCNet获得。
{"title":"Spatial transcriptomics expression prediction from histopathology based on cross-modal mask reconstruction and contrastive learning","authors":"Junzhuo Liu ,&nbsp;Markus Eckstein ,&nbsp;Zhixiang Wang ,&nbsp;Friedrich Feuerhake ,&nbsp;Dorit Merhof","doi":"10.1016/j.media.2025.103889","DOIUrl":"10.1016/j.media.2025.103889","url":null,"abstract":"<div><div>Spatial transcriptomics is a technology that captures gene expression at different spatial locations, widely used in tumor microenvironment analysis and molecular profiling of histopathology, providing valuable insights into resolving gene expression and clinical diagnosis of cancer. Due to the high cost of data acquisition, large-scale spatial transcriptomics data remain challenging to obtain. In this study, we develop a contrastive learning-based deep learning method to predict spatially resolved gene expression from the whole-slide images (WSIs). Unlike existing end-to-end prediction frameworks, our method leverages multi-modal contrastive learning to establish a correspondence between histopathological morphology and spatial gene expression in the feature space. By computing cross-modal feature similarity, our method generates spatially resolved gene expression directly from WSIs. Furthermore, to enhance the standard contrastive learning paradigm, a cross-modal masked reconstruction is designed as a pretext task, enabling feature-level fusion between modalities. Notably, our method does not rely on large-scale pretraining datasets or abstract semantic representations from either modality, making it particularly effective for scenarios with limited spatial transcriptomics data. Evaluation across six different disease datasets demonstrates that, compared to existing studies, our method improves Pearson Correlation Coefficient (PCC) in the prediction of highly expressed genes, highly variable genes, and marker genes by 6.27 %, 6.11 %, and 11.26 % respectively. Further analysis indicates that our method preserves gene-gene correlations and applies to datasets with limited samples. Additionally, our method exhibits potential in cancer tissue localization based on biomarker expression. The code repository for this work is available at <span><span>https://github.com/ngfufdrdh/CMRCNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"108 ","pages":"Article 103889"},"PeriodicalIF":11.8,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145592833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised learning of spatially varying regularization for diffeomorphic image registration 差分图像配准中空间变化正则化的无监督学习
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-24 DOI: 10.1016/j.media.2025.103887
Junyu Chen , Shuwen Wei , Yihao Liu , Zhangxing Bian , Yufan He , Aaron Carass , Harrison Bai , Yong Du
Spatially varying regularization accommodates the deformation variations that may be necessary for different anatomical regions during deformable image registration. Historically, optimization-based registration models have harnessed spatially varying regularization to address anatomical subtleties. However, most modern deep learning-based models tend to gravitate towards spatially invariant regularization, wherein a homogenous regularization strength is applied across the entire image, potentially disregarding localized variations. In this paper, we propose a hierarchical probabilistic model that integrates a prior distribution on the deformation regularization strength, enabling the end-to-end learning of a spatially varying deformation regularizer directly from the data. The proposed method is straightforward to implement and easily integrates with various registration network architectures. Additionally, automatic tuning of hyperparameters is achieved through Bayesian optimization, allowing efficient identification of optimal hyperparameters for any given registration task. Comprehensive evaluations on publicly available datasets demonstrate that the proposed method significantly improves registration performance and enhances the interpretability of deep learning-based registration, all while maintaining smooth deformations. Our code is freely available at http://bit.ly/3BrXGxz.
空间变化正则化适应在可变形图像配准期间不同解剖区域可能需要的变形变化。从历史上看,基于优化的配准模型利用空间变化的正则化来解决解剖学的微妙之处。然而,大多数现代基于深度学习的模型倾向于空间不变正则化,其中在整个图像上应用均匀的正则化强度,可能忽略局部变化。在本文中,我们提出了一个分层概率模型,该模型集成了变形正则化强度的先验分布,从而可以直接从数据中端到端学习空间变化的变形正则化器。该方法实现简单,易于与各种注册网络体系结构集成。此外,超参数的自动调优是通过贝叶斯优化实现的,允许有效地识别任何给定配准任务的最优超参数。对公开可用数据集的综合评估表明,该方法显著提高了配准性能,增强了基于深度学习的配准的可解释性,同时保持了平滑的形变。我们的代码可以在http://bit.ly/3BrXGxz上免费获得。
{"title":"Unsupervised learning of spatially varying regularization for diffeomorphic image registration","authors":"Junyu Chen ,&nbsp;Shuwen Wei ,&nbsp;Yihao Liu ,&nbsp;Zhangxing Bian ,&nbsp;Yufan He ,&nbsp;Aaron Carass ,&nbsp;Harrison Bai ,&nbsp;Yong Du","doi":"10.1016/j.media.2025.103887","DOIUrl":"10.1016/j.media.2025.103887","url":null,"abstract":"<div><div>Spatially varying regularization accommodates the deformation variations that may be necessary for different anatomical regions during deformable image registration. Historically, optimization-based registration models have harnessed spatially varying regularization to address anatomical subtleties. However, most modern deep learning-based models tend to gravitate towards spatially invariant regularization, wherein a homogenous regularization strength is applied across the entire image, potentially disregarding localized variations. In this paper, we propose a hierarchical probabilistic model that integrates a prior distribution on the deformation regularization strength, enabling the end-to-end learning of a spatially varying deformation regularizer directly from the data. The proposed method is straightforward to implement and easily integrates with various registration network architectures. Additionally, automatic tuning of hyperparameters is achieved through Bayesian optimization, allowing efficient identification of optimal hyperparameters for any given registration task. Comprehensive evaluations on publicly available datasets demonstrate that the proposed method significantly improves registration performance and enhances the interpretability of deep learning-based registration, all while maintaining smooth deformations. Our code is freely available at <span><span>http://bit.ly/3BrXGxz</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"108 ","pages":"Article 103887"},"PeriodicalIF":11.8,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145592834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Extreme cardiac MRI analysis under respiratory motion: Results of the CMRxMotion challenge 呼吸运动下的极端心脏MRI分析:CMRxMotion挑战的结果
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-22 DOI: 10.1016/j.media.2025.103883
Kang Wang , Chen Qin , Zhang Shi , Haoran Wang , Xiwen Zhang , Chen Chen , Cheng Ouyang , Chengliang Dai , Yuanhan Mo , Chenchen Dai , Xutong Kuang , Ruizhe Li , Xin Chen , Xiuzheng Yue , Song Tian , Alejandro Mora-Rubio , Kumaradevan Punithakumar , Shizhan Gong , Qi Dou , Sina Amirrajab , Shuo Wang
Deep learning models have achieved state-of-the-art performance in automated Cardiac Magnetic Resonance (CMR) analysis. However, the efficacy of these models is highly dependent on the availability of high-quality, artifact-free images. In clinical practice, CMR acquisitions are frequently degraded by respiratory motion, yet the robustness of deep learning models against such artifacts remains an underexplored problem. To promote research in this domain, we organized the MICCAI CMRxMotion challenge. We curated and publicly released a dataset of 320 CMR cine series from 40 healthy volunteers who performed specific breathing protocols to induce a controlled spectrum of motion artifacts. The challenge comprised two tasks: 1) automated image quality assessment to classify images based on motion severity, and 2) robust myocardial segmentation in the presence of motion artifacts. A total of 22 algorithms were submitted and evaluated on the two designated tasks. This paper presents a comprehensive overview of the challenge design and dataset, reports the evaluation results for the top-performing methods, and further investigates the impact of motion artifacts on five clinically relevant biomarkers. All resources and code are publicly available at: https://github.com/CMRxMotion.
深度学习模型在自动心脏磁共振(CMR)分析中取得了最先进的性能。然而,这些模型的有效性高度依赖于高质量,无伪像图像的可用性。在临床实践中,CMR采集经常因呼吸运动而退化,然而深度学习模型对这些伪像的鲁棒性仍然是一个未被充分探索的问题。为了促进这一领域的研究,我们组织了MICCAI CMRxMotion挑战赛。我们策划并公开发布了来自40名健康志愿者的320个CMR电影系列的数据集,这些志愿者执行特定的呼吸方案来诱导受控的运动伪影频谱。该挑战包括两个任务:1)自动图像质量评估,根据运动严重程度对图像进行分类;2)存在运动伪影的鲁棒心肌分割。针对这两个指定的任务,总共提交了22种算法并进行了评估。本文全面概述了挑战设计和数据集,报告了表现最好的方法的评估结果,并进一步研究了运动伪影对五种临床相关生物标志物的影响。所有的资源和代码都可以在https://github.com/CMRxMotion上公开获得。
{"title":"Extreme cardiac MRI analysis under respiratory motion: Results of the CMRxMotion challenge","authors":"Kang Wang ,&nbsp;Chen Qin ,&nbsp;Zhang Shi ,&nbsp;Haoran Wang ,&nbsp;Xiwen Zhang ,&nbsp;Chen Chen ,&nbsp;Cheng Ouyang ,&nbsp;Chengliang Dai ,&nbsp;Yuanhan Mo ,&nbsp;Chenchen Dai ,&nbsp;Xutong Kuang ,&nbsp;Ruizhe Li ,&nbsp;Xin Chen ,&nbsp;Xiuzheng Yue ,&nbsp;Song Tian ,&nbsp;Alejandro Mora-Rubio ,&nbsp;Kumaradevan Punithakumar ,&nbsp;Shizhan Gong ,&nbsp;Qi Dou ,&nbsp;Sina Amirrajab ,&nbsp;Shuo Wang","doi":"10.1016/j.media.2025.103883","DOIUrl":"10.1016/j.media.2025.103883","url":null,"abstract":"<div><div>Deep learning models have achieved state-of-the-art performance in automated Cardiac Magnetic Resonance (CMR) analysis. However, the efficacy of these models is highly dependent on the availability of high-quality, artifact-free images. In clinical practice, CMR acquisitions are frequently degraded by respiratory motion, yet the robustness of deep learning models against such artifacts remains an underexplored problem. To promote research in this domain, we organized the <em>MICCAI CMRxMotion</em> challenge. We curated and publicly released a dataset of 320 CMR cine series from 40 healthy volunteers who performed specific breathing protocols to induce a controlled spectrum of motion artifacts. The challenge comprised two tasks: 1) automated image quality assessment to classify images based on motion severity, and 2) robust myocardial segmentation in the presence of motion artifacts. A total of 22 algorithms were submitted and evaluated on the two designated tasks. This paper presents a comprehensive overview of the challenge design and dataset, reports the evaluation results for the top-performing methods, and further investigates the impact of motion artifacts on five clinically relevant biomarkers. All resources and code are publicly available at: <span><span>https://github.com/CMRxMotion</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103883"},"PeriodicalIF":11.8,"publicationDate":"2025-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145567299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PL-Seg: Partially labeled abdominal organ segmentation via classwise orthogonal contrastive learning and progressive self-distillation l - seg:通过分类正交对比学习和渐进式自蒸馏进行部分标记腹部器官分割
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-21 DOI: 10.1016/j.media.2025.103885
He Li , Xiangde Luo , Jia Fu , Ran Gu , Wenjun Liao , Shichuan Zhang , Kang Li , Guotai Wang , Shaoting Zhang
Accurate segmentation of abdominal organs in Computed Tomography (CT) scans is crucial for effective lesion diagnosis, radiotherapy planning, and patient follow-up. Although deep learning has shown great performance with fully supervised learning, it requires voxel-level dense annotations that are time-consuming and costly to obtain, especially for multiple organs. In this work, we propose a novel framework PL-Seg, for multi-organ segmentation in abdominal CT scans using partially labeled data, where only a subset of the target organ classes are annotated in each volume. First, we introduce a novel Hardness-Aware Decoupled Foreground Loss (HADFL), which focuses exclusively on annotated organs and dynamically adjusts class weights based on historical segmentation difficulty. Then, we employ a Classwise Orthogonal Contrastive Loss (COCL) to reduce inter-class ambiguity, which serves as a regularization for unlabeled regions. In addition, a Progressive Self-Distillation (PSD) that distills knowledge from deep high-resolution layers to shallower low-resolution levels is proposed to improve the feature learning ability under partial class annotations. Experiments conducted on a dataset with varying class-wise annotation ratios and a real clinical partially labeled dataset demonstrate that: 1) PL-Seg achieves significant performance improvements by leveraging unlabeled categories, 2) Compared with six state-of-the-art methods, PL-Seg achieves superior results with a simpler pipeline and greater computational efficiency, and 3) Under the same annotation cost, PL-Seg outperforms existing semi-supervised methods. Furthermore, we release a partially labeled medical image segmentation codebase and benchmark to boost research on this topic: https://github.com/HiLab-git/PLS4MIS.
计算机断层扫描(CT)中腹部器官的准确分割对于有效的病变诊断,放疗计划和患者随访至关重要。尽管深度学习在完全监督学习中表现出了很好的性能,但它需要体素级的密集注释,这是耗时和昂贵的,特别是对于多器官。在这项工作中,我们提出了一个新的框架PL-Seg,用于腹部CT扫描中使用部分标记数据的多器官分割,其中每个卷中只有目标器官类别的一个子集被注释。首先,我们引入了一种新的硬度感知解耦前景损失算法(HADFL),该算法只关注标注器官,并根据历史分割难度动态调整类权值。然后,我们使用分类正交对比损失(COCL)来减少类间歧义,这可以作为未标记区域的正则化。此外,为了提高部分类标注下的特征学习能力,提出了一种递进式自蒸馏(Progressive Self-Distillation, PSD)方法,将知识从深度高分辨率层提炼到较浅的低分辨率层。在不同类别标注比例的数据集和真实临床部分标注数据集上进行的实验表明:1)PL-Seg通过利用未标注的类别获得了显著的性能提升;2)与六种最先进的方法相比,PL-Seg以更简单的管道和更高的计算效率获得了更优的结果;3)在相同的标注成本下,PL-Seg优于现有的半监督方法。此外,我们发布了一个部分标记的医学图像分割代码库和基准,以促进对该主题的研究:https://github.com/HiLab-git/PLS4MIS。
{"title":"PL-Seg: Partially labeled abdominal organ segmentation via classwise orthogonal contrastive learning and progressive self-distillation","authors":"He Li ,&nbsp;Xiangde Luo ,&nbsp;Jia Fu ,&nbsp;Ran Gu ,&nbsp;Wenjun Liao ,&nbsp;Shichuan Zhang ,&nbsp;Kang Li ,&nbsp;Guotai Wang ,&nbsp;Shaoting Zhang","doi":"10.1016/j.media.2025.103885","DOIUrl":"10.1016/j.media.2025.103885","url":null,"abstract":"<div><div>Accurate segmentation of abdominal organs in Computed Tomography (CT) scans is crucial for effective lesion diagnosis, radiotherapy planning, and patient follow-up. Although deep learning has shown great performance with fully supervised learning, it requires voxel-level dense annotations that are time-consuming and costly to obtain, especially for multiple organs. In this work, we propose a novel framework PL-Seg, for multi-organ segmentation in abdominal CT scans using partially labeled data, where only a subset of the target organ classes are annotated in each volume. First, we introduce a novel Hardness-Aware Decoupled Foreground Loss (HADFL), which focuses exclusively on annotated organs and dynamically adjusts class weights based on historical segmentation difficulty. Then, we employ a Classwise Orthogonal Contrastive Loss (COCL) to reduce inter-class ambiguity, which serves as a regularization for unlabeled regions. In addition, a Progressive Self-Distillation (PSD) that distills knowledge from deep high-resolution layers to shallower low-resolution levels is proposed to improve the feature learning ability under partial class annotations. Experiments conducted on a dataset with varying class-wise annotation ratios and a real clinical partially labeled dataset demonstrate that: 1) PL-Seg achieves significant performance improvements by leveraging unlabeled categories, 2) Compared with six state-of-the-art methods, PL-Seg achieves superior results with a simpler pipeline and greater computational efficiency, and 3) Under the same annotation cost, PL-Seg outperforms existing semi-supervised methods. Furthermore, we release a partially labeled medical image segmentation codebase and benchmark to boost research on this topic: <span><span>https://github.com/HiLab-git/PLS4MIS</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"108 ","pages":"Article 103885"},"PeriodicalIF":11.8,"publicationDate":"2025-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145568010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Direction-Aware convolution for airway tubular feature enhancement network 气道管状特征增强网络的方向感知卷积
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-20 DOI: 10.1016/j.media.2025.103882
Qibiao Wu , Yagang Wang , Qian Zhang
Manual annotation of airway regions in computed tomography images is a time-consuming and expertise-dependent task. Automatic airway segmentation is therefore a prerequisite for enabling rapid bronchoscopic navigation and the clinical deployment of bronchoscopic robotic systems. Although convolutional neural network methods have gained considerable attention in airway segmentation, the unique tree-like structure of airways poses challenges for conventional and deformable convolutions, which often fail to focus on fine airway structures, leading to missed segments and discontinuities. To address this issue, this study proposes a novel tubular feature extraction network, named TfeNet. TfeNet introduces a novel direction-aware convolution operator that adapts the geometry of linear convolution kernels through spatial rotation transformations, enabling it to dynamically align with the tubular structures of airways and effectively enhance feature extraction. Furthermore, a tubular feature fusion module (TFFM) is designed based on asymmetric convolution and residual connection strategies, effectively capturing the features of airway tubules from different directions. Extensive experiments conducted on one public dataset and two datasets used in airway segmentation challenges demonstrate the effectiveness of TfeNet. Specifically, our method achieves a comprehensive lead in both accuracy and continuity on the BAS dataset, attains the highest mean score of 94.95 % on the ATM22 dataset by balancing accuracy and continuity, and demonstrates superior leakage control and precision on the challenging AIIB23 dataset. The code is available at https://github.com/QibiaoWu/TfeNet.
在计算机断层扫描图像中手动标注气道区域是一项耗时且依赖专业知识的任务。因此,自动气道分割是实现快速支气管镜导航和支气管镜机器人系统临床部署的先决条件。尽管卷积神经网络方法在气道分割中得到了相当多的关注,但气道独特的树状结构给传统的可变形卷积带来了挑战,传统卷积往往不能关注精细的气道结构,导致遗漏的部分和不连续。为了解决这个问题,本研究提出了一种新的管状特征提取网络,命名为TfeNet。TfeNet引入了一种新的方向感知卷积算子,该算子通过空间旋转变换来适应线性卷积核的几何形状,使其能够与气道管状结构动态对齐,有效增强特征提取。在此基础上,设计了基于非对称卷积和残差连接策略的气道小管特征融合模块(TFFM),从不同方向有效捕获气道小管特征。在一个公共数据集和两个用于气道分割挑战的数据集上进行的大量实验证明了TfeNet的有效性。具体而言,我们的方法在BAS数据集的精度和连续性方面都取得了全面领先,在ATM22数据集上达到了94.95%的最高平均分数,在具有挑战性的AIIB23数据集上表现出了卓越的泄漏控制和精度。代码可在https://github.com/QibiaoWu/TfeNet上获得。
{"title":"Direction-Aware convolution for airway tubular feature enhancement network","authors":"Qibiao Wu ,&nbsp;Yagang Wang ,&nbsp;Qian Zhang","doi":"10.1016/j.media.2025.103882","DOIUrl":"10.1016/j.media.2025.103882","url":null,"abstract":"<div><div>Manual annotation of airway regions in computed tomography images is a time-consuming and expertise-dependent task. Automatic airway segmentation is therefore a prerequisite for enabling rapid bronchoscopic navigation and the clinical deployment of bronchoscopic robotic systems. Although convolutional neural network methods have gained considerable attention in airway segmentation, the unique tree-like structure of airways poses challenges for conventional and deformable convolutions, which often fail to focus on fine airway structures, leading to missed segments and discontinuities. To address this issue, this study proposes a novel tubular feature extraction network, named TfeNet. TfeNet introduces a novel direction-aware convolution operator that adapts the geometry of linear convolution kernels through spatial rotation transformations, enabling it to dynamically align with the tubular structures of airways and effectively enhance feature extraction. Furthermore, a tubular feature fusion module (TFFM) is designed based on asymmetric convolution and residual connection strategies, effectively capturing the features of airway tubules from different directions. Extensive experiments conducted on one public dataset and two datasets used in airway segmentation challenges demonstrate the effectiveness of TfeNet. Specifically, our method achieves a comprehensive lead in both accuracy and continuity on the BAS dataset, attains the highest mean score of 94.95 % on the ATM22 dataset by balancing accuracy and continuity, and demonstrates superior leakage control and precision on the challenging AIIB23 dataset. The code is available at <span><span>https://github.com/QibiaoWu/TfeNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"108 ","pages":"Article 103882"},"PeriodicalIF":11.8,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145559818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A systematic analysis of the impact of data variation on AI-based histopathological grading of prostate cancer 数据变化对基于人工智能的前列腺癌组织病理分级影响的系统分析
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-20 DOI: 10.1016/j.media.2025.103884
Patrick Fuhlert , Fabian Westhaeusser , Esther Dietrich , Maximilian Lennartz , Robin Khatri , Nico Kaiser , Pontus Röbeck , Roman Bülow , Saskia Von Stillfried , Anja Witte , Sam Ladjevardi , Anders Drotte , Peter Severgårdh , Jan Baumbach , Victor G. Puelles , Michael Häggman , Michael Brehler , Peter Boor , Peter Walhagen , Anca Dragomir , Stefan Bonn
The histopathological evaluation of biopsies by human experts is a gold standard in clinical disease diagnosis. While recent artificial intelligence-based (AI) approaches have reached human expert-level performance, they often display shortcomings caused by variations in sample preparation, limiting clinical applicability. This study investigates the impact of data variation on AI-based histopathological grading and explores algorithmic approaches that confer prediction robustness. To evaluate the impact of data variation in histopathology, we collected a multicentric, retrospective, observational prostate cancer (PCa) trial consisting of six cohorts in 3 countries with 25,591 patients, 83,864 images. This includes a high-variance dataset of 8,157 patients and 28,236 images with variations in section thickness, staining protocol, and scanner. This unique training dataset enabled the development of an AI-based PCa grading framework by training on patient outcome, not subjective grading. It was made robust through several algorithmic adaptations, including domain adversarial training and credibility-guided color adaptation. We named the final grading framework PCAI. We compare PCAI to a BASE model and human experts on three external test cohorts, comprising 2,255 patients and 9,437 images. Variations in sample processing, particularly section thickness and staining time, significantly reduced the performance of AI-based PCa grading by up to 8.6 percentage points in the event-ordered concordance index (EOC-Index) thus highlighting serious risks for AI-based histopathological grading. Algorithmic improvements for model robustness, credibility, and training on high-variance data as well as outcome-based severity prediction give rise to robust models with grading performance surpassing experienced pathologists. We demonstrate how our algorithmic enhancements for greater robustness lead to significantly better performance, surpassing expert grading on EOC-Index and 5-year AUROC by up to 21.2 percentage points.
人类专家对活组织检查的组织病理学评估是临床疾病诊断的金标准。虽然最近基于人工智能(AI)的方法已经达到了人类专家水平的表现,但它们经常显示出由样品制备变化引起的缺点,限制了临床适用性。本研究调查了数据变化对基于人工智能的组织病理学分级的影响,并探索了赋予预测鲁棒性的算法方法。为了评估组织病理学数据变化的影响,我们收集了一项多中心,回顾性,观察性前列腺癌(PCa)试验,包括3个国家的6个队列,25,591例患者,83,864张图像。这包括8,157名患者和28,236张不同切片厚度、染色方案和扫描仪的图像的高方差数据集。这个独特的训练数据集通过对患者结果的训练,而不是主观评分,实现了基于人工智能的PCa评分框架的开发。通过几个算法的调整,包括领域对抗训练和可信度引导的颜色适应,它变得鲁棒。我们将最终的评分框架命名为PCAI。我们将PCAI与BASE模型和人类专家在三个外部测试队列中进行比较,其中包括2,255名患者和9,437张图像。样品处理的变化,特别是切片厚度和染色时间的变化,显著降低了基于人工智能的PCa分级的性能,其事件顺序一致性指数(EOC-Index)高达8.6个百分点,从而突出了基于人工智能的组织病理学分级的严重风险。对模型鲁棒性、可信度和高方差数据训练的算法改进,以及基于结果的严重程度预测,产生了鲁棒模型,其分级性能超过了经验丰富的病理学家。我们展示了我们的算法增强如何提高鲁棒性,从而显著提高性能,超过专家对EOC-Index和5年AUROC的评分高达21.2个百分点。
{"title":"A systematic analysis of the impact of data variation on AI-based histopathological grading of prostate cancer","authors":"Patrick Fuhlert ,&nbsp;Fabian Westhaeusser ,&nbsp;Esther Dietrich ,&nbsp;Maximilian Lennartz ,&nbsp;Robin Khatri ,&nbsp;Nico Kaiser ,&nbsp;Pontus Röbeck ,&nbsp;Roman Bülow ,&nbsp;Saskia Von Stillfried ,&nbsp;Anja Witte ,&nbsp;Sam Ladjevardi ,&nbsp;Anders Drotte ,&nbsp;Peter Severgårdh ,&nbsp;Jan Baumbach ,&nbsp;Victor G. Puelles ,&nbsp;Michael Häggman ,&nbsp;Michael Brehler ,&nbsp;Peter Boor ,&nbsp;Peter Walhagen ,&nbsp;Anca Dragomir ,&nbsp;Stefan Bonn","doi":"10.1016/j.media.2025.103884","DOIUrl":"10.1016/j.media.2025.103884","url":null,"abstract":"<div><div>The histopathological evaluation of biopsies by human experts is a gold standard in clinical disease diagnosis. While recent artificial intelligence-based (AI) approaches have reached human expert-level performance, they often display shortcomings caused by variations in sample preparation, limiting clinical applicability. This study investigates the impact of data variation on AI-based histopathological grading and explores algorithmic approaches that confer prediction robustness. To evaluate the impact of data variation in histopathology, we collected a multicentric, retrospective, observational prostate cancer (PCa) trial consisting of six cohorts in 3 countries with 25,591 patients, 83,864 images. This includes a high-variance dataset of 8,157 patients and 28,236 images with variations in section thickness, staining protocol, and scanner. This unique training dataset enabled the development of an AI-based PCa grading framework by training on patient outcome, not subjective grading. It was made robust through several algorithmic adaptations, including domain adversarial training and credibility-guided color adaptation. We named the final grading framework PCAI. We compare PCAI to a BASE model and human experts on three external test cohorts, comprising 2,255 patients and 9,437 images. Variations in sample processing, particularly section thickness and staining time, significantly reduced the performance of AI-based PCa grading by up to 8.6 percentage points in the event-ordered concordance index (EOC-Index) thus highlighting serious risks for AI-based histopathological grading. Algorithmic improvements for model robustness, credibility, and training on high-variance data as well as outcome-based severity prediction give rise to robust models with grading performance surpassing experienced pathologists. We demonstrate how our algorithmic enhancements for greater robustness lead to significantly better performance, surpassing expert grading on EOC-Index and 5-year AUROC by up to 21.2 percentage points.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"108 ","pages":"Article 103884"},"PeriodicalIF":11.8,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145559819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Openness-aware multi-prototype learning for open set medical diagnosis 开放集医学诊断的开放感知多原型学习
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-19 DOI: 10.1016/j.media.2025.103863
Mingyuan Liu , Lu Xu , Yuzhuo Gu , Jicong Zhang , Shuo Li
Unlike the prevalent image classification paradigm that assumes all samples belong to pre-defined classes, Open set recognition (OSR) indicates that new classes unobserved during training could appear in testing. It mandates a model to not only categorize known classes but also recognize unknowns. Existing prototype-based solutions model each class using a single prototype and recognize samples that are distant from these prototypes as unknowns. However, single-prototype modeling overlooks intra-class variance, leading to large open space risk. Additionally, open space regularization is ignored, allowing unknown samples to remain in their initial positions that overlap with the known space, thus impeding unknown discrimination. To address these limitations, we propose Openness-Aware Multi-Prototype Learning (OAMPL) with two novel designs: (1) Adaptive Open Multi-Prototype Formulation (AOMF) extends single-prototype modeling to a novel multi-prototype formulation. It reduces open space risk by simultaneously avoiding class underrepresentation and anticipating unknown occurrences. Additionally, AOMF incorporates a balancing term, a marginal factor, and a learnable scalar to flexibly fit intricate open environments. (2) Difficulty Aware Openness Simulator (DAOS) dynamically synthesizes fake features at varying difficulties to represent open classes. By punishing the adjacency between the fake and the known, the known-unknown discrimination could be enhanced. DAOS is distinguished by its joint optimization with AOMF, allowing it to cooperate with the classifier to produce samples with appropriate difficulties for effective learning. As OSR is a nascent topic in medical fields, we contribute three benchmark datasets. Compared with state-of-the-art models, our OAMPL maintains closed set accuracy and achieves improvements in OSR at about 1.5 % and 1.2 % measured by AUROC and OSCR, respectively. Extensive ablation experiments demonstrate the effectiveness of each design.
与假设所有样本都属于预定义类的流行图像分类范式不同,开放集识别(OSR)表明在训练过程中未观察到的新类可能出现在测试中。它要求模型不仅要对已知类进行分类,还要识别未知类。现有的基于原型的解决方案使用单个原型对每个类建模,并将远离这些原型的样本识别为未知。然而,单原型建模忽略了类内方差,导致很大的开放空间风险。此外,开放空间正则化被忽略,允许未知样本保持在与已知空间重叠的初始位置,从而阻碍未知判别。为了解决这些限制,我们提出了两种新颖设计的开放感知多原型学习(OAMPL):(1)自适应开放多原型公式(AOMF)将单原型建模扩展到新的多原型公式。它通过同时避免阶级代表性不足和预测未知事件来降低开放空间风险。此外,AOMF还结合了一个平衡项、一个边缘因子和一个可学习标量,以灵活地适应复杂的开放环境。(2)难度感知开放模拟器(DAOS)动态合成不同难度的假特征来表示开放类。通过对已知与虚假之间的邻接性进行惩罚,可以增强已知与未知的区分。DAOS的特点是与AOMF联合优化,使其能够与分类器合作产生适当难度的样本以进行有效的学习。由于OSR在医学领域是一个新兴的话题,我们提供了三个基准数据集。与最先进的模型相比,我们的OAMPL保持了闭集精度,OSR分别提高了1.5%和1.2%,分别由AUROC和OSCR测量。大量的烧蚀实验证明了每种设计的有效性。
{"title":"Openness-aware multi-prototype learning for open set medical diagnosis","authors":"Mingyuan Liu ,&nbsp;Lu Xu ,&nbsp;Yuzhuo Gu ,&nbsp;Jicong Zhang ,&nbsp;Shuo Li","doi":"10.1016/j.media.2025.103863","DOIUrl":"10.1016/j.media.2025.103863","url":null,"abstract":"<div><div>Unlike the prevalent image classification paradigm that assumes all samples belong to pre-defined classes, Open set recognition (OSR) indicates that new classes unobserved during training could appear in testing. It mandates a model to not only categorize known classes but also recognize unknowns. Existing prototype-based solutions model each class using a single prototype and recognize samples that are distant from these prototypes as unknowns. However, single-prototype modeling overlooks intra-class variance, leading to large open space risk. Additionally, open space regularization is ignored, allowing unknown samples to remain in their initial positions that overlap with the known space, thus impeding unknown discrimination. To address these limitations, we propose Openness-Aware Multi-Prototype Learning (OAMPL) with two novel designs: (1) Adaptive Open Multi-Prototype Formulation (AOMF) extends single-prototype modeling to a novel multi-prototype formulation. It reduces open space risk by simultaneously avoiding class underrepresentation and anticipating unknown occurrences. Additionally, AOMF incorporates a balancing term, a marginal factor, and a learnable scalar to flexibly fit intricate open environments. (2) Difficulty Aware Openness Simulator (DAOS) dynamically synthesizes fake features at varying difficulties to represent open classes. By punishing the adjacency between the fake and the known, the known-unknown discrimination could be enhanced. DAOS is distinguished by its joint optimization with AOMF, allowing it to cooperate with the classifier to produce samples with appropriate difficulties for effective learning. As OSR is a nascent topic in medical fields, we contribute three benchmark datasets. Compared with state-of-the-art models, our OAMPL maintains closed set accuracy and achieves improvements in OSR at about 1.5 % and 1.2 % measured by AUROC and OSCR, respectively. Extensive ablation experiments demonstrate the effectiveness of each design.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"108 ","pages":"Article 103863"},"PeriodicalIF":11.8,"publicationDate":"2025-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145553607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Multi-instance Learning Network with Prototype-instance Adversarial Contrastive for Cervix Pathology Grading 基于原型-实例对抗对比的多实例学习网络用于子宫颈病理分级
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-17 DOI: 10.1016/j.media.2025.103880
Mingrui Ma , Furong Luo , Binlin Ma , Shuxian Liu , Xiaoyi Lv , Pan Huang
The pathological grading of cervical squamous cell carcinoma (CSCC) is a fundamental and important index in tumor diagnosis. Pathologists tend to focus on single differentiation areas during the grading process. Existing multi-instance learning (MIL) methods divide pathology images into regions, generating multiple differentiated instances (MDIs) that often exhibit ambiguous grading patterns. These ambiguities reduce the model’s ability to accurately represent CSCC pathological grading patterns. Motivated by these issues, we propose an end-to-end multi-instance learning network with prototype-instance adversarial contrastive learning, termed PacMIL, which incorporates three key ideas. First, we introduce an end-to-end multi-instance nonequilibrium learning algorithm that addresses the mismatch between MIL feature representations and CSCC pathological grading, and enables nonequilibrium representation. Second, we design a prototype-instance adversarial contrastive (PAC) approach that integrates a priori prototype instances and a probability distribution attention mechanism. This enhances the model’s ability to learn representations for single differentiated instances (SDIs). Third, we incorporate an adversarial contrastive learning strategy into the PAC method to overcome the limitation that fixed metrics rarely capture the variability of MDIs and SDIs. In addition, we embed the correct metric distances of the MDIs and SDIs into the optimization objective function to further guide representation learning. Extensive experiments demonstrate that our PacMIL model achieves 93.09% and 0.9802 for the mAcc and AUC metrics, respectively, outperforming other SOTA models. Moreover, the representation ability of PacMIL is superior to that of existing SOTA approaches. Overall, our model offers enhanced practicality in CSCC pathological grading. Our code and dataset will be publicly available at https://github.com/Baron-Huang/PacMIL.
宫颈鳞状细胞癌(CSCC)的病理分级是诊断宫颈鳞状细胞癌的基础和重要指标。病理学家在分级过程中倾向于关注单一的分化区域。现有的多实例学习(MIL)方法将病理图像划分为多个区域,产生的多分化实例(mdi)往往表现出模糊的分级模式。这些模糊性降低了模型准确代表CSCC病理分级模式的能力。基于这些问题,我们提出了一个包含原型-实例对抗对比学习的端到端多实例学习网络,称为PacMIL,它包含三个关键思想。首先,我们引入了一种端到端的多实例非平衡学习算法,该算法解决了MIL特征表示与CSCC病理分级之间的不匹配问题,并实现了非平衡表示。其次,我们设计了一种原型-实例对抗对比(PAC)方法,该方法集成了先验原型实例和概率分布注意机制。这增强了模型学习单差异化实例(sdi)表示的能力。第三,我们将对抗性对比学习策略纳入PAC方法,以克服固定指标很少捕获mdi和sdi可变性的限制。此外,我们将mdi和sdi的正确度量距离嵌入到优化目标函数中,以进一步指导表示学习。大量的实验表明,我们的PacMIL模型在mAcc和AUC指标上分别达到了93.09%和0.9802,优于其他SOTA模型。此外,PacMIL的表示能力优于现有的SOTA方法。总之,我们的模型增强了CSCC病理分级的实用性。我们的代码和数据集将在https://github.com/Baron-Huang/PacMIL上公开提供。
{"title":"A Multi-instance Learning Network with Prototype-instance Adversarial Contrastive for Cervix Pathology Grading","authors":"Mingrui Ma ,&nbsp;Furong Luo ,&nbsp;Binlin Ma ,&nbsp;Shuxian Liu ,&nbsp;Xiaoyi Lv ,&nbsp;Pan Huang","doi":"10.1016/j.media.2025.103880","DOIUrl":"10.1016/j.media.2025.103880","url":null,"abstract":"<div><div>The pathological grading of cervical squamous cell carcinoma (CSCC) is a fundamental and important index in tumor diagnosis. Pathologists tend to focus on single differentiation areas during the grading process. Existing multi-instance learning (MIL) methods divide pathology images into regions, generating multiple differentiated instances (MDIs) that often exhibit ambiguous grading patterns. These ambiguities reduce the model’s ability to accurately represent CSCC pathological grading patterns. Motivated by these issues, we propose an end-to-end multi-instance learning network with prototype-instance adversarial contrastive learning, termed PacMIL, which incorporates three key ideas. First, we introduce an end-to-end multi-instance nonequilibrium learning algorithm that addresses the mismatch between MIL feature representations and CSCC pathological grading, and enables nonequilibrium representation. Second, we design a prototype-instance adversarial contrastive (PAC) approach that integrates a priori prototype instances and a probability distribution attention mechanism. This enhances the model’s ability to learn representations for single differentiated instances (SDIs). Third, we incorporate an adversarial contrastive learning strategy into the PAC method to overcome the limitation that fixed metrics rarely capture the variability of MDIs and SDIs. In addition, we embed the correct metric distances of the MDIs and SDIs into the optimization objective function to further guide representation learning. Extensive experiments demonstrate that our PacMIL model achieves 93.09% and 0.9802 for the mAcc and AUC metrics, respectively, outperforming other SOTA models. Moreover, the representation ability of PacMIL is superior to that of existing SOTA approaches. Overall, our model offers enhanced practicality in CSCC pathological grading. Our code and dataset will be publicly available at <span><span><em>https://github.com/Baron-Huang/PacMIL</em></span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103880"},"PeriodicalIF":11.8,"publicationDate":"2025-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145535684","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Medical image analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1