首页 > 最新文献

Medical image analysis最新文献

英文 中文
Diagnostic text-guided representation learning in hierarchical classification for pathological whole slide image 诊断性文本引导表征学习在病理整片图像分层分类中的应用
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-07 DOI: 10.1016/j.media.2025.103894
Jiawen Li , Qiehe Sun , Renao Yan , Yizhi Wang , Yuqiu Fu , Yani Wei , Tian Guan , Huijuan Shi , Yonghong He , Anjia Han
With the development of digital imaging in medical microscopy, artificial intelligent-based analysis of pathological whole slide images (WSIs) provides a powerful tool for cancer diagnosis. Limited by the expensive cost of pixel-level annotation, current research primarily focuses on representation learning with slide-level labels, showing success in various downstream tasks. However, given the diversity of lesion types and the complex relationships between each other, these techniques still deserve further exploration in addressing advanced pathology tasks. To this end, we introduce the concept of hierarchical pathological image classification and propose a representation learning called PathTree. PathTree considers the multi-classification of diseases as a binary tree structure. Each category is represented as a professional pathological text description, which messages information with a tree-like encoder. The interactive text features are then used to guide the aggregation of hierarchical multiple representations. PathTree uses slide-text similarity to obtain probability scores and introduces two extra tree-specific losses to further constrain the association between texts and slides. Through extensive experiments on three challenging hierarchical classification datasets: in-house cryosectioned lung tissue lesion identification, public prostate cancer grade assessment, and public breast cancer subtyping, our proposed PathTree is consistently competitive compared to the state-of-the-art methods and provides a new perspective on the deep learning-assisted solution for more complex WSI classification.
随着数字成像技术在医学显微镜中的发展,基于人工智能的病理全切片图像分析为癌症诊断提供了有力的工具。受像素级标注昂贵成本的限制,目前的研究主要集中在使用幻灯片级标签的表示学习上,并在各种下游任务中取得了成功。然而,鉴于病变类型的多样性和彼此之间的复杂关系,这些技术在解决高级病理任务方面仍值得进一步探索。为此,我们引入了分层病理图像分类的概念,并提出了一种称为PathTree的表示学习方法。PathTree将疾病的多重分类视为二叉树结构。每个类别都表示为一个专业的病理文本描述,它用树状编码器传递信息。然后使用交互式文本特征来指导分层多表示的聚合。PathTree使用幻灯片-文本相似性来获得概率分数,并引入两个额外的树特定损失来进一步约束文本和幻灯片之间的关联。通过在三个具有挑战性的分层分类数据集上的广泛实验:内部冷冻切片肺组织病变识别、公共前列腺癌分级评估和公共乳腺癌亚型,我们提出的PathTree与最先进的方法相比始终具有竞争力,并为更复杂的WSI分类提供了深度学习辅助解决方案的新视角。
{"title":"Diagnostic text-guided representation learning in hierarchical classification for pathological whole slide image","authors":"Jiawen Li ,&nbsp;Qiehe Sun ,&nbsp;Renao Yan ,&nbsp;Yizhi Wang ,&nbsp;Yuqiu Fu ,&nbsp;Yani Wei ,&nbsp;Tian Guan ,&nbsp;Huijuan Shi ,&nbsp;Yonghong He ,&nbsp;Anjia Han","doi":"10.1016/j.media.2025.103894","DOIUrl":"10.1016/j.media.2025.103894","url":null,"abstract":"<div><div>With the development of digital imaging in medical microscopy, artificial intelligent-based analysis of pathological whole slide images (WSIs) provides a powerful tool for cancer diagnosis. Limited by the expensive cost of pixel-level annotation, current research primarily focuses on representation learning with slide-level labels, showing success in various downstream tasks. However, given the diversity of lesion types and the complex relationships between each other, these techniques still deserve further exploration in addressing advanced pathology tasks. To this end, we introduce the concept of hierarchical pathological image classification and propose a representation learning called PathTree. PathTree considers the multi-classification of diseases as a binary tree structure. Each category is represented as a professional pathological text description, which messages information with a tree-like encoder. The interactive text features are then used to guide the aggregation of hierarchical multiple representations. PathTree uses slide-text similarity to obtain probability scores and introduces two extra tree-specific losses to further constrain the association between texts and slides. Through extensive experiments on three challenging hierarchical classification datasets: in-house cryosectioned lung tissue lesion identification, public prostate cancer grade assessment, and public breast cancer subtyping, our proposed PathTree is consistently competitive compared to the state-of-the-art methods and provides a new perspective on the deep learning-assisted solution for more complex WSI classification.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103894"},"PeriodicalIF":11.8,"publicationDate":"2025-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145689528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Immunocto: A massive immune cell database auto-generated for histopathology 免疫细胞:为组织病理学自动生成的大量免疫细胞数据库
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-07 DOI: 10.1016/j.media.2025.103905
Mikaël Simard , Zhuoyan Shen , Konstantin Bräutigam , Rasha Abu-Eid , Maria A. Hawkins , Charles-Antoine Collins Fekete
With the advent of novel cancer treatment options such as immunotherapy, studying the tumour immune micro-environment (TIME) is crucial to inform on prognosis and understand potential response to therapeutic agents. A key approach to characterising the TIME involves combining digitised images of haematoxylin and eosin (H&E) stained tissue sections obtained in routine histopathology examination with automated immune cell detection and classification methods. In this work, we introduce a workflow to automatically generate robust single cell contours and labels from dually stained tissue sections with H&E and multiplexed immunofluorescence (IF) markers. The approach harnesses the Segment Anything Model and requires minimal human intervention compared to existing single cell databases. With this methodology, we create Immunocto, a massive, multi-million automatically generated database of 6,848,454 human cells and objects, including 2,282,818 immune cells distributed across 4 subtypes: CD4+ T cell lymphocytes, CD8+ T cell lymphocytes, CD20+ B cell lymphocytes, and CD68+/CD163+ macrophages. For each cell, we provide a 64 × 64 pixels2 H&E image at 40 ×  magnification, along with a binary mask of the nucleus and a label. The database, which is made publicly available, can be used to train models to study the TIME on routine H&E slides. We show that deep learning models trained on Immunocto result in state-of-the-art performance for lymphocyte detection. The approach demonstrates the benefits of using matched H&E and IF data to generate robust databases for computational pathology applications.
随着免疫疗法等新型癌症治疗方案的出现,研究肿瘤免疫微环境(TIME)对于预测预后和了解治疗药物的潜在反应至关重要。表征TIME的关键方法包括将常规组织病理学检查中获得的血红素和伊红(H&;E)染色组织切片的数字化图像与自动免疫细胞检测和分类方法相结合。在这项工作中,我们引入了一个工作流程,可以自动生成具有H&;E和多重免疫荧光(IF)标记的双染色组织切片的鲁棒单细胞轮廓和标签。该方法利用分段任意模型,与现有的单细胞数据库相比,需要最少的人为干预。利用这种方法,我们创建了Immunocto,这是一个巨大的、数百万自动生成的数据库,包含6848454个人类细胞和物体,包括2282818个免疫细胞,分布在4个亚型:CD4+ T细胞淋巴细胞、CD8+ T细胞淋巴细胞、CD20+ B细胞淋巴细胞和CD68+/CD163+巨噬细胞。对于每个细胞,我们提供了一张64 × 64 pixels2的H&;E图像,放大倍数为40 ×,以及细胞核的二进制掩模和标签。该数据库是公开的,可用于训练模型来研究常规H&;E幻灯片上的TIME。我们表明,在Immunocto上训练的深度学习模型在淋巴细胞检测方面具有最先进的性能。该方法证明了使用匹配的H&;E和IF数据为计算病理学应用生成健壮的数据库的好处。
{"title":"Immunocto: A massive immune cell database auto-generated for histopathology","authors":"Mikaël Simard ,&nbsp;Zhuoyan Shen ,&nbsp;Konstantin Bräutigam ,&nbsp;Rasha Abu-Eid ,&nbsp;Maria A. Hawkins ,&nbsp;Charles-Antoine Collins Fekete","doi":"10.1016/j.media.2025.103905","DOIUrl":"10.1016/j.media.2025.103905","url":null,"abstract":"<div><div>With the advent of novel cancer treatment options such as immunotherapy, studying the tumour immune micro-environment (TIME) is crucial to inform on prognosis and understand potential response to therapeutic agents. A key approach to characterising the TIME involves combining digitised images of haematoxylin and eosin (H&amp;E) stained tissue sections obtained in routine histopathology examination with automated immune cell detection and classification methods. In this work, we introduce a workflow to automatically generate robust single cell contours and labels from dually stained tissue sections with H&amp;E and multiplexed immunofluorescence (IF) markers. The approach harnesses the Segment Anything Model and requires minimal human intervention compared to existing single cell databases. With this methodology, we create Immunocto, a massive, multi-million automatically generated database of 6,848,454 human cells and objects, including 2,282,818 immune cells distributed across 4 subtypes: CD4<span><math><msup><mrow></mrow><mo>+</mo></msup></math></span> T cell lymphocytes, CD8<span><math><msup><mrow></mrow><mo>+</mo></msup></math></span> T cell lymphocytes, CD20<span><math><msup><mrow></mrow><mo>+</mo></msup></math></span> B cell lymphocytes, and CD68<span><math><msup><mrow></mrow><mo>+</mo></msup></math></span>/CD163<span><math><msup><mrow></mrow><mo>+</mo></msup></math></span> macrophages. For each cell, we provide a 64 × 64 pixels<sup>2</sup> H&amp;E image at <strong>40</strong> ×  magnification, along with a binary mask of the nucleus and a label. The database, which is made publicly available, can be used to train models to study the TIME on routine H&amp;E slides. We show that deep learning models trained on Immunocto result in state-of-the-art performance for lymphocyte detection. The approach demonstrates the benefits of using matched H&amp;E and IF data to generate robust databases for computational pathology applications.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103905"},"PeriodicalIF":11.8,"publicationDate":"2025-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145705011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Perivascular space identification nnUNet for generalised usage (PINGU) 通用血管周围空间识别网(PINGU)
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-04 DOI: 10.1016/j.media.2025.103903
Benjamin Sinclair , William Pham , Lucy Vivash , Jasmine Moses , Miranda Lynch , Karina Dorfman , Cassandra Marotta , Shaun Koh , Jacob Bunyamin , Ella Rowsthorn , Alex Jarema , Himashi Peiris , Zhaolin Chen , Sandy R Shultz , David K Wright , Dexiao Kong , Sharon L. Naismith , Terence J. O’Brien , Meng Law
Perivascular spaces (PVSs) form a central component of the brain’s waste clearance system, the glymphatic system. These structures are visible on MRIs when enlarged, and their morphology is associated with aging and neurological disease. Manual quantification of PVS is time consuming and subjective. Numerous deep learning methods for PVS segmentation have been developed for automated segmentation. However, the majority of these algorithms have been developed and evaluated on homogenous datasets and high resolution scans, perhaps limiting their applicability for the wide range of image qualities acquired in clinical and research settings. In this work we train a nnUNet, a top-performing task driven biomedical image segmentation deep learning algorithm, on a heterogenous training sample of manually segmented MRIs of a range of different qualities and resolutions from 7 different datasets acquired on 6 different scanners. These are compared to the two currently publicly available deep learning methods for 3D segmentation of PVS, evaluated on scans with a range of resolutions and qualities. The resulting model, PINGU (Perivascular space Identification Nnunet for Generalised Usage), achieved voxel and cluster level dice scores of 0.50(SD=0.15) and 0.63(0.17) in the white matter (WM), and 0.54 (0.11) and 0.66(0.17) in the basal ganglia (BG). Performance on unseen “external” sites’ data was substantially lower for both PINGU (0.20-0.38 [WM, voxel], 0.29-0.58 [WM, cluster], 0.22-0.36 [BG, voxel], 0.46-0.60 [BG, cluster]) and the publicly available algorithms (0.18-0.30 [WM, voxel], 0.29-0.38 [WM cluster], 0.10-0.20 [BG, voxel], 0.15-0.37 [BG, cluster]). Nonetheless, PINGU strongly outperformed the publicly available algorithms, particularly in the BG. PINGU stands out as broad-use PVS segmentation tool, with particular strength in the BG, an area of PVS highly related to vascular disease and pathology.
血管周围空间(PVSs)是大脑废物清除系统(淋巴系统)的中心组成部分。这些结构放大后在mri上可见,其形态与衰老和神经系统疾病有关。人工量化pv既耗时又主观。为了实现自动分割,已经开发了许多用于pv分割的深度学习方法。然而,这些算法中的大多数都是在同质数据集和高分辨率扫描上开发和评估的,这可能限制了它们在临床和研究环境中获得的广泛图像质量的适用性。在这项工作中,我们训练了nnUNet,一种性能最好的任务驱动的生物医学图像分割深度学习算法,在来自6个不同扫描仪上获得的7个不同数据集的一系列不同质量和分辨率的手动分割mri的异构训练样本上。将这些方法与目前公开的两种用于pv 3D分割的深度学习方法进行比较,并对扫描结果进行一系列分辨率和质量的评估。所得模型PINGU(血管周围空间识别Nnunet for Generalised Usage)在白质(WM)中获得了0.50(SD=0.15)和0.63(0.17)的体素和聚类水平骰子分数,在基底节区(BG)中获得了0.54(0.11)和0.66(0.17)的分数。PINGU (0.20-0.38 [WM,体素],0.29-0.58 [WM,聚类],0.22-0.36 [BG,体素],0.46-0.60 [BG,聚类])和公开的算法(0.18-0.30 [WM,体素],0.29-0.38 [WM聚类],0.10-0.20 [BG,体素],0.15-0.37 [BG,聚类])在未见过的“外部”站点数据上的性能都明显较低。尽管如此,PINGU的表现仍然远远优于公开可用的算法,特别是在BG中。PINGU作为一种广泛使用的PVS分割工具,在与血管疾病和病理高度相关的PVS领域BG中具有特别的优势。
{"title":"Perivascular space identification nnUNet for generalised usage (PINGU)","authors":"Benjamin Sinclair ,&nbsp;William Pham ,&nbsp;Lucy Vivash ,&nbsp;Jasmine Moses ,&nbsp;Miranda Lynch ,&nbsp;Karina Dorfman ,&nbsp;Cassandra Marotta ,&nbsp;Shaun Koh ,&nbsp;Jacob Bunyamin ,&nbsp;Ella Rowsthorn ,&nbsp;Alex Jarema ,&nbsp;Himashi Peiris ,&nbsp;Zhaolin Chen ,&nbsp;Sandy R Shultz ,&nbsp;David K Wright ,&nbsp;Dexiao Kong ,&nbsp;Sharon L. Naismith ,&nbsp;Terence J. O’Brien ,&nbsp;Meng Law","doi":"10.1016/j.media.2025.103903","DOIUrl":"10.1016/j.media.2025.103903","url":null,"abstract":"<div><div>Perivascular spaces (PVSs) form a central component of the brain’s waste clearance system, the glymphatic system. These structures are visible on MRIs when enlarged, and their morphology is associated with aging and neurological disease. Manual quantification of PVS is time consuming and subjective. Numerous deep learning methods for PVS segmentation have been developed for automated segmentation. However, the majority of these algorithms have been developed and evaluated on homogenous datasets and high resolution scans, perhaps limiting their applicability for the wide range of image qualities acquired in clinical and research settings. In this work we train a nnUNet, a top-performing task driven biomedical image segmentation deep learning algorithm, on a heterogenous training sample of manually segmented MRIs of a range of different qualities and resolutions from 7 different datasets acquired on 6 different scanners. These are compared to the two currently publicly available deep learning methods for 3D segmentation of PVS, evaluated on scans with a range of resolutions and qualities. The resulting model, PINGU (Perivascular space Identification Nnunet for Generalised Usage), achieved voxel and cluster level dice scores of 0.50(SD=0.15) and 0.63(0.17) in the white matter (WM), and 0.54 (0.11) and 0.66(0.17) in the basal ganglia (BG). Performance on unseen “external” sites’ data was substantially lower for both PINGU (0.20-0.38 [WM, voxel], 0.29-0.58 [WM, cluster], 0.22-0.36 [BG, voxel], 0.46-0.60 [BG, cluster]) and the publicly available algorithms (0.18-0.30 [WM, voxel], 0.29-0.38 [WM cluster], 0.10-0.20 [BG, voxel], 0.15-0.37 [BG, cluster]). Nonetheless, PINGU strongly outperformed the publicly available algorithms, particularly in the BG. PINGU stands out as broad-use PVS segmentation tool, with particular strength in the BG, an area of PVS highly related to vascular disease and pathology.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103903"},"PeriodicalIF":11.8,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145689529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AGFS-tractometry: A novel atlas-guided fine-scale tractometry approach for enhanced along-tract group statistical comparison using diffusion MRI tractography agfs - tractetry:一种新的阿特拉斯引导的精细尺度tractetry方法,用于增强沿束组统计比较,使用扩散MRI Tractography
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-03 DOI: 10.1016/j.media.2025.103892
Ruixi Zheng , Wei Zhang , Yijie Li , Xi Zhu , Zhou Lan , Jarrett Rushmore , Yogesh Rathi , Nikos Makris , Lauren J. O’Donnell , Fan Zhang
Diffusion MRI (dMRI) tractography is currently the only method for in vivo mapping of the brain’s white matter (WM) connections. Tractometry is an advanced tractography analysis technique for along-tract profiling to investigate the morphology and microstructural properties along the fiber tracts. Tractometry has become an essential tool for studying local along-tract differences between different populations (e.g., health vs disease). In this study, we propose a novel atlas-guided fine-scale tractometry method, namely AGFS-Tractometry, that leverages tract spatial information and permutation testing to enhance the along-tract statistical analysis between populations. There are two major contributions in AGFS-Tractometry. First, we create a novel atlas-guided tract profiling template that enables consistent, fine-scale, along-tract parcellation of subject-specific fiber tracts. Second, we propose a novel nonparametric permutation testing group comparison method to enable simultaneous analysis across all along-tract parcels while correcting for multiple comparisons. We perform experimental evaluations on synthetic datasets with known group differences and in vivo real data. We compare AGFS-Tractometry with two state-of-the-art tractometry methods, including Automated Fiber-tract Quantification (AFQ) and BUndle ANalytics (BUAN). Our results show that the proposed AGFS-Tractometry obtains enhanced sensitivity and specificity in detecting local WM differences. In the real data analysis experiments, AGFS-Tractometry can identify more regions with significant differences, which are anatomically consistent with the existing literature. Overall, these demonstrate the ability of AGFS-Tractometry to detect subtle or spatially localized WM group-level differences. The created tract profiling template and related code are available at: https://github.com/ZhengRuixi/AGFS-Tractometry.git.
扩散磁共振成像(dMRI)神经束成像是目前唯一的在体脑白质(WM)连接成像方法。纤维束法是一种先进的纤维束分析技术,用于研究纤维束的形态和微观结构特性。束测法已成为研究不同人群之间局部沿束差异(例如,健康vs疾病)的重要工具。在本研究中,我们提出了一种新的阿特拉斯引导的精细尺度tractometry方法,即AGFS-Tractometry,该方法利用通道空间信息和排列测试来增强种群间的沿通道统计分析。在AGFS-Tractometry中有两个主要贡献。首先,我们创建了一种新的图谱引导的纤维束分析模板,该模板可以实现一致的、精细的、沿着特定纤维束的束状分割。其次,我们提出了一种新的非参数排列测试组比较方法,以便在校正多重比较的同时对所有沿道包裹进行同时分析。我们对已知组差异和体内真实数据的合成数据集进行实验评估。我们将AGFS-Tractometry与两种最先进的tractometry方法进行比较,包括自动纤维束定量(AFQ)和束分析(BUAN)。我们的研究结果表明,所提出的AGFS-Tractometry在检测局部WM差异方面具有更高的灵敏度和特异性。在真实的数据分析实验中,AGFS-Tractometry能够识别出更多具有显著差异的区域,这在解剖学上与已有文献一致。总的来说,这些表明AGFS-Tractometry能够检测细微的或空间定位的WM组水平差异。创建的通道分析模板和相关代码可从https://github.com/ZhengRuixi/AGFS-Tractometry.git获得。
{"title":"AGFS-tractometry: A novel atlas-guided fine-scale tractometry approach for enhanced along-tract group statistical comparison using diffusion MRI tractography","authors":"Ruixi Zheng ,&nbsp;Wei Zhang ,&nbsp;Yijie Li ,&nbsp;Xi Zhu ,&nbsp;Zhou Lan ,&nbsp;Jarrett Rushmore ,&nbsp;Yogesh Rathi ,&nbsp;Nikos Makris ,&nbsp;Lauren J. O’Donnell ,&nbsp;Fan Zhang","doi":"10.1016/j.media.2025.103892","DOIUrl":"10.1016/j.media.2025.103892","url":null,"abstract":"<div><div>Diffusion MRI (dMRI) tractography is currently the only method for in vivo mapping of the brain’s white matter (WM) connections. Tractometry is an advanced tractography analysis technique for along-tract profiling to investigate the morphology and microstructural properties along the fiber tracts. Tractometry has become an essential tool for studying local along-tract differences between different populations (e.g., health vs disease). In this study, we propose a novel atlas-guided fine-scale tractometry method, namely AGFS-Tractometry, that leverages tract spatial information and permutation testing to enhance the along-tract statistical analysis between populations. There are two major contributions in AGFS-Tractometry. First, we create a novel atlas-guided tract profiling template that enables consistent, fine-scale, along-tract parcellation of subject-specific fiber tracts. Second, we propose a novel nonparametric permutation testing group comparison method to enable simultaneous analysis across all along-tract parcels while correcting for multiple comparisons. We perform experimental evaluations on synthetic datasets with known group differences and in vivo real data. We compare AGFS-Tractometry with two state-of-the-art tractometry methods, including Automated Fiber-tract Quantification (AFQ) and BUndle ANalytics (BUAN). Our results show that the proposed AGFS-Tractometry obtains enhanced sensitivity and specificity in detecting local WM differences. In the real data analysis experiments, AGFS-Tractometry can identify more regions with significant differences, which are anatomically consistent with the existing literature. Overall, these demonstrate the ability of AGFS-Tractometry to detect subtle or spatially localized WM group-level differences. The created tract profiling template and related code are available at: <span><span>https://github.com/ZhengRuixi/AGFS-Tractometry.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103892"},"PeriodicalIF":11.8,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145689531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncertainty estimates in pharmacokinetic modelling of DCE-MRI DCE-MRI药代动力学模型的不确定性估计
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-03 DOI: 10.1016/j.media.2025.103881
Jonas M. Van Elburg , Natalia V. Korobova , Mohammad M. Islam , Marian A. Troelstra , Oliver J. Gurney-Champion
Dynamic contrast-enhanced (DCE) MRI is a powerful technique for detecting and characterising various diseases by quantifying tissue perfusion. However, accurate perfusion quantification remains challenging due to noisy data and the complexity of pharmacokinetic modelling. Conventional non-linear least squares (NLLS) fitting often yields noisy parameter maps. Although deep-learning algorithms generate smoother, more visually appealing maps, these may lure clinicians into a false sense of security when the maps are incorrect. Hence, reliable uncertainty estimation is crucial for assessing model performance and ensuring clinical confidence.
Therefore, we implemented an ensemble of mean-variance estimation (MVE) neural networks to quantify perfusion parameters alongside aleatoric (data-driven) and epistemic (model-driven) uncertainties in DCE-MRI. We compared MVE with NLLS and a physics-informed neural network (PINN), both of which used conventional covariance matrix-based uncertainty estimation.
Simulations demonstrated that MVE achieved the highest accuracy in perfusion and uncertainty estimates. MVE’s aleatoric uncertainty strongly correlated with true errors, whereas NLLS and PINN tended to overestimate uncertainty. Epistemic uncertainty was significantly higher for the data deviating from what was encountered in training (out-of-distribution) in both MVE and PINN ensembles. In vivo, MVE produced smoother and more reliable uncertainty maps than NLLS and PINN, which exhibited outliers and overestimation. Within a liver region of interest, MVE’s uncertainty estimates matched the standard deviation of the data more closely than NLLS and PINN, making it the most accurate method.
In conclusion, an MVE enhances quantitative DCE-MRI by providing robust uncertainty estimates alongside perfusion parameters. This approach improves the reliability of AI-driven MRI analysis, supporting clinical translation.
动态对比增强(DCE) MRI是一种通过定量组织灌注来检测和表征各种疾病的强大技术。然而,由于数据嘈杂和药代动力学建模的复杂性,准确的灌注定量仍然具有挑战性。传统的非线性最小二乘(NLLS)拟合通常会产生噪声参数图。尽管深度学习算法生成的地图更平滑,视觉上更吸引人,但当地图不正确时,这些算法可能会诱使临床医生产生错误的安全感。因此,可靠的不确定性估计对于评估模型性能和确保临床置信度至关重要。因此,我们实现了均值方差估计(MVE)神经网络的集合,以量化灌注参数以及DCE-MRI中的任意(数据驱动)和认知(模型驱动)不确定性。我们将MVE与NLLS和物理信息神经网络(PINN)进行了比较,这两种方法都使用传统的基于协方差矩阵的不确定性估计。仿真结果表明,MVE在灌注和不确定性估计方面具有最高的准确性。MVE的任意不确定性与真实误差密切相关,而NLLS和PINN倾向于高估不确定性。在MVE和PINN集成中,对于偏离训练中遇到的数据(分布外),认知不确定性明显更高。在体内,与NLLS和PINN相比,MVE产生的不确定性图更平滑、更可靠,后者表现出异常值和高估。在感兴趣的肝脏区域内,MVE的不确定性估计比NLLS和PINN更接近数据的标准偏差,使其成为最准确的方法。总之,MVE通过提供可靠的不确定性估计和灌注参数来增强定量DCE-MRI。这种方法提高了人工智能驱动的MRI分析的可靠性,支持临床翻译。
{"title":"Uncertainty estimates in pharmacokinetic modelling of DCE-MRI","authors":"Jonas M. Van Elburg ,&nbsp;Natalia V. Korobova ,&nbsp;Mohammad M. Islam ,&nbsp;Marian A. Troelstra ,&nbsp;Oliver J. Gurney-Champion","doi":"10.1016/j.media.2025.103881","DOIUrl":"10.1016/j.media.2025.103881","url":null,"abstract":"<div><div>Dynamic contrast-enhanced (DCE) MRI is a powerful technique for detecting and characterising various diseases by quantifying tissue perfusion. However, accurate perfusion quantification remains challenging due to noisy data and the complexity of pharmacokinetic modelling. Conventional non-linear least squares (NLLS) fitting often yields noisy parameter maps. Although deep-learning algorithms generate smoother, more visually appealing maps, these may lure clinicians into a false sense of security when the maps are incorrect. Hence, reliable uncertainty estimation is crucial for assessing model performance and ensuring clinical confidence.</div><div>Therefore, we implemented an ensemble of mean-variance estimation (MVE) neural networks to quantify perfusion parameters alongside aleatoric (data-driven) and epistemic (model-driven) uncertainties in DCE-MRI. We compared MVE with NLLS and a physics-informed neural network (PINN), both of which used conventional covariance matrix-based uncertainty estimation.</div><div>Simulations demonstrated that MVE achieved the highest accuracy in perfusion and uncertainty estimates. MVE’s aleatoric uncertainty strongly correlated with true errors, whereas NLLS and PINN tended to overestimate uncertainty. Epistemic uncertainty was significantly higher for the data deviating from what was encountered in training (out-of-distribution) in both MVE and PINN ensembles. In vivo, MVE produced smoother and more reliable uncertainty maps than NLLS and PINN, which exhibited outliers and overestimation. Within a liver region of interest, MVE’s uncertainty estimates matched the standard deviation of the data more closely than NLLS and PINN, making it the most accurate method.</div><div>In conclusion, an MVE enhances quantitative DCE-MRI by providing robust uncertainty estimates alongside perfusion parameters. This approach improves the reliability of AI-driven MRI analysis, supporting clinical translation.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103881"},"PeriodicalIF":11.8,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145689530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
D-EDL: Differential evidential deep learning for robust medical out-of-distribution detection D-EDL:基于差分证据深度学习的鲁棒医学非分布检测
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-30 DOI: 10.1016/j.media.2025.103888
Wei Fu , Yufei Chen , Yuqi Liu , Xiaodong Yue
In computer-aided diagnosis, the extreme imbalance in disease incidence rates often results in the omission of rare conditions, leading to out-of-distribution (OOD) samples during testing. To prevent unreliable diagnostic outputs, detecting these OOD samples becomes essential for clinical jnsafety. While Evidential Deep Learning (EDL) and its variants have shown great promise in detecting outliers, their clinical application remains challenging due to the variability in medical images. We find that when encountering samples with high data uncertainty, the Kullback-Leibler divergence (KL) in EDL tends to suppress inherent ambiguity, resulting in an over-penalty effect in evidence estimation that impairs discrimination between ambiguous in-distribution cases and true outliers. Motivated by the confirmatory and differential diagnostic process in clinical practice, we propose Differential Evidential Deep Learning (D-EDL), a simple but effective method for robust OOD detection. Specifically, we treat KL as a confirmatory restriction and innovatively replace it with a Ruling Out Module (ROM) for differential restriction, which reduces over-penalty on ambiguous ID samples while maintaining OOD sensitivity. Considering extreme testing scenarios, we introduce test-time Raw evidence Inference (RI) to bypass instability in uncertainty estimation with refined evidence and further improve robustness and precision. Finally, we propose the Balanced Detection Score (BDS) to quantify the potential on clinical performance when optimally balancing misdiagnoses and missed diagnoses across varying sensitivities. Experimental results on ISIC2019, Bone Marrow Cytomorphology datasets and EDDFS dataset demonstrate that our D-EDL outperforms state-of-the-art OOD detection methods, achieving significant improvements in robustness and clinical applicability. Code for D-EDL is available at https://github.com/KellaDoe/Differential_EDL.
在计算机辅助诊断中,由于疾病发病率的极度不平衡,往往会导致罕见疾病的遗漏,从而导致检测过程中的样本分布不均匀(OOD)。为了防止诊断结果不可靠,检测这些OOD样本对于临床安全至关重要。虽然证据深度学习(EDL)及其变体在检测异常值方面显示出巨大的希望,但由于医学图像的可变性,它们的临床应用仍然具有挑战性。我们发现,当遇到具有高数据不确定性的样本时,EDL中的Kullback-Leibler散度(KL)倾向于抑制固有的模糊性,从而导致证据估计中的过度惩罚效应,损害了分布中模糊性案例与真实异常值之间的区分。在临床实践的验证性和鉴别诊断过程的推动下,我们提出了差分证据深度学习(D-EDL),这是一种简单但有效的鲁棒性OOD检测方法。具体而言,我们将KL视为验证性限制,并创新地将其替换为排除模块(ROM)以进行差分限制,这减少了对模糊ID样本的过度惩罚,同时保持了OOD灵敏度。考虑到极端测试场景,我们引入了测试时间原始证据推断(RI)来绕过不确定性估计中的不稳定性,进一步提高了鲁棒性和精度。最后,我们提出了平衡检测评分(BDS)来量化在不同敏感性之间最佳平衡误诊和漏诊时临床表现的潜力。在ISIC2019、骨髓细胞形态学数据集和EDDFS数据集上的实验结果表明,我们的D-EDL优于最先进的OOD检测方法,在稳健性和临床适用性方面取得了显着提高。D-EDL的代码可从https://github.com/KellaDoe/Differential_EDL获得。
{"title":"D-EDL: Differential evidential deep learning for robust medical out-of-distribution detection","authors":"Wei Fu ,&nbsp;Yufei Chen ,&nbsp;Yuqi Liu ,&nbsp;Xiaodong Yue","doi":"10.1016/j.media.2025.103888","DOIUrl":"10.1016/j.media.2025.103888","url":null,"abstract":"<div><div>In computer-aided diagnosis, the extreme imbalance in disease incidence rates often results in the omission of rare conditions, leading to out-of-distribution (OOD) samples during testing. To prevent unreliable diagnostic outputs, detecting these OOD samples becomes essential for clinical jnsafety. While Evidential Deep Learning (EDL) and its variants have shown great promise in detecting outliers, their clinical application remains challenging due to the variability in medical images. We find that when encountering samples with high data uncertainty, the Kullback-Leibler divergence (KL) in EDL tends to suppress inherent ambiguity, resulting in an over-penalty effect in evidence estimation that impairs discrimination between ambiguous in-distribution cases and true outliers. Motivated by the confirmatory and differential diagnostic process in clinical practice, we propose Differential Evidential Deep Learning (D-EDL), a simple but effective method for robust OOD detection. Specifically, we treat KL as a confirmatory restriction and innovatively replace it with a Ruling Out Module (ROM) for differential restriction, which reduces over-penalty on ambiguous ID samples while maintaining OOD sensitivity. Considering extreme testing scenarios, we introduce test-time Raw evidence Inference (RI) to bypass instability in uncertainty estimation with refined evidence and further improve robustness and precision. Finally, we propose the Balanced Detection Score (BDS) to quantify the potential on clinical performance when optimally balancing misdiagnoses and missed diagnoses across varying sensitivities. Experimental results on ISIC2019, Bone Marrow Cytomorphology datasets and EDDFS dataset demonstrate that our D-EDL outperforms state-of-the-art OOD detection methods, achieving significant improvements in robustness and clinical applicability. Code for D-EDL is available at <span><span>https://github.com/KellaDoe/Differential_EDL</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103888"},"PeriodicalIF":11.8,"publicationDate":"2025-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145619565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PISCO: Self-supervised k-space regularization for improved neural implicit k-space representations of dynamic MRI 动态MRI神经隐式k空间表征的自监督k空间正则化
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-29 DOI: 10.1016/j.media.2025.103890
Veronika Spieker , Hannah Eichhorn , Wenqi Huang , Jonathan K. Stelter , Tabita Catalan , Rickmer F. Braren , Daniel Rueckert , Francisco Sahli Costabal , Kerstin Hammernik , Dimitrios C. Karampinos , Claudia Prieto , Julia A. Schnabel
Neural implicit k-space representations (NIK) have shown promising results for dynamic magnetic resonance imaging (MRI) at high temporal resolutions. Yet, reducing acquisition time, and thereby available training data, results in severe performance drops due to overfitting. To address this, we introduce a novel self-supervised k-space loss function LPISCO, applicable for regularization of NIK-based reconstructions. The proposed loss function is based on the concept of parallel imaging-inspired self-consistency (PISCO), enforcing a consistent global k-space neighborhood relationship without requiring additional data. Quantitative and qualitative evaluations on static and dynamic MR reconstructions show that integrating PISCO significantly improves NIK representations, making it a competitive dynamic reconstruction method without constraining the temporal resolution. Particularly at high acceleration factors (R ≥ 50), NIK with PISCO can avoid temporal oversmoothing of state-of-the-art methods and achieves superior spatio-temporal reconstruction quality. Furthermore, an extensive analysis of the loss assumptions and stability shows PISCO’s potential as versatile self-supervised k-space loss function for further applications and architectures. Code is available at: https://github.com/compai-lab/2025-pisco-spieker
神经隐式k空间表示(NIK)在高时间分辨率下的动态磁共振成像(MRI)中显示出有希望的结果。然而,减少获取时间,从而减少可用的训练数据,会导致过度拟合导致严重的性能下降。为了解决这个问题,我们引入了一种新的自监督k空间损失函数LPISCO,适用于基于nik的重构的正则化。所提出的损失函数基于并行成像启发自洽(PISCO)的概念,在不需要额外数据的情况下强制执行一致的全局k空间邻域关系。静态和动态MR重建的定量和定性评价表明,整合PISCO显著改善了NIK表征,使其成为一种具有竞争力的动态重建方法,而不限制时间分辨率。特别是在高加速度因子(R ≥ 50)下,采用PISCO的NIK可以避免现有方法的时间过平滑,获得较好的时空重建质量。此外,对损失假设和稳定性的广泛分析表明,PISCO作为多功能自监督k空间损失函数的潜力可用于进一步的应用和体系结构。代码可从https://github.com/compai-lab/2025-pisco-spieker获得
{"title":"PISCO: Self-supervised k-space regularization for improved neural implicit k-space representations of dynamic MRI","authors":"Veronika Spieker ,&nbsp;Hannah Eichhorn ,&nbsp;Wenqi Huang ,&nbsp;Jonathan K. Stelter ,&nbsp;Tabita Catalan ,&nbsp;Rickmer F. Braren ,&nbsp;Daniel Rueckert ,&nbsp;Francisco Sahli Costabal ,&nbsp;Kerstin Hammernik ,&nbsp;Dimitrios C. Karampinos ,&nbsp;Claudia Prieto ,&nbsp;Julia A. Schnabel","doi":"10.1016/j.media.2025.103890","DOIUrl":"10.1016/j.media.2025.103890","url":null,"abstract":"<div><div>Neural implicit k-space representations (NIK) have shown promising results for dynamic magnetic resonance imaging (MRI) at high temporal resolutions. Yet, reducing acquisition time, and thereby available training data, results in severe performance drops due to overfitting. To address this, we introduce a novel self-supervised k-space loss function <span><math><msub><mi>L</mi><mtext>PISCO</mtext></msub></math></span>, applicable for regularization of NIK-based reconstructions. The proposed loss function is based on the concept of parallel imaging-inspired self-consistency (PISCO), enforcing a consistent global k-space neighborhood relationship without requiring additional data. Quantitative and qualitative evaluations on static and dynamic MR reconstructions show that integrating PISCO significantly improves NIK representations, making it a competitive dynamic reconstruction method without constraining the temporal resolution. Particularly at high acceleration factors (R ≥ 50), NIK with PISCO can avoid temporal oversmoothing of state-of-the-art methods and achieves superior spatio-temporal reconstruction quality. Furthermore, an extensive analysis of the loss assumptions and stability shows PISCO’s potential as versatile self-supervised k-space loss function for further applications and architectures. Code is available at: <span><span>https://github.com/compai-lab/2025-pisco-spieker</span><svg><path></path></svg></span></div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103890"},"PeriodicalIF":11.8,"publicationDate":"2025-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145619566","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GEM-pRF: GPU-empowered mapping of population receptive fields for large-scale fMRI analysis GEM-pRF:基于gpu的大规模功能磁共振成像(fMRI)分析的群体接受域映射
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-28 DOI: 10.1016/j.media.2025.103891
Siddharth Mittal, Michael Woletz, David Linhardt, Christian Windischberger
Population receptive field (pRF) mapping is a fundamental technique for understanding retinotopic organisation of the human visual system. Since its introduction in 2008, however, its scalability has been severely hindered by the computational bottleneck of iterative parameter refinement. Current state-of-the-art implementations either sacrifice precision for speed or rely on slow iterative parameter updates, limiting their applicability to large-scale datasets. Here, we present a novel mathematical reformulation of the General Linear Model (GLM), wrapped in a GPU-Empowered Mapping of population Receptive Fields (GEM-pRF) software implementation. By orthogonalizing the design matrix, our approach enables the direct and fast computation of the objective function’s derivatives, which are used to eliminate the iterative refinement process. This approach dramatically accelerates pRF estimation with high accuracy. Validation using empirical and simulated data confirms GEM-pRF’s accuracy, and benchmarking against established tools demonstrates a reduction in computation time of almost two orders of magnitude. With its modular and extensible design, GEM-pRF provides a critical advancement for large-scale fMRI retinotopic mapping. Furthermore, our reformulated GLM approach in combination with GPU-based implementation offers a broadly applicable solution that may extend beyond visual neuroscience, accelerating computational modelling across various domains in neuroimaging and beyond.
群体感受野(pRF)映射是理解人类视觉系统视网膜组织的基本技术。然而,自2008年推出以来,其可扩展性受到迭代参数细化的计算瓶颈的严重阻碍。当前最先进的实现要么牺牲精度换取速度,要么依赖缓慢的迭代参数更新,限制了它们对大规模数据集的适用性。在这里,我们提出了一个新的通用线性模型(GLM)的数学重新表述,包裹在一个gpu授权的群体接受域映射(GEM-pRF)软件实现中。通过正交化设计矩阵,我们的方法可以直接和快速地计算目标函数的导数,从而消除迭代优化过程。该方法极大地提高了pRF估计的精度。使用经验和模拟数据的验证证实了GEM-pRF的准确性,针对既定工具的基准测试表明,计算时间减少了近两个数量级。由于其模块化和可扩展的设计,GEM-pRF为大规模功能磁共振成像视网膜定位提供了关键的进步。此外,我们重新制定的GLM方法与基于gpu的实现相结合,提供了一种广泛适用的解决方案,可以扩展到视觉神经科学之外,加速神经成像等各个领域的计算建模。
{"title":"GEM-pRF: GPU-empowered mapping of population receptive fields for large-scale fMRI analysis","authors":"Siddharth Mittal,&nbsp;Michael Woletz,&nbsp;David Linhardt,&nbsp;Christian Windischberger","doi":"10.1016/j.media.2025.103891","DOIUrl":"10.1016/j.media.2025.103891","url":null,"abstract":"<div><div>Population receptive field (pRF) mapping is a fundamental technique for understanding retinotopic organisation of the human visual system. Since its introduction in 2008, however, its scalability has been severely hindered by the computational bottleneck of iterative parameter refinement. Current state-of-the-art implementations either sacrifice precision for speed or rely on slow iterative parameter updates, limiting their applicability to large-scale datasets. Here, we present a novel mathematical reformulation of the General Linear Model (GLM), wrapped in a GPU-Empowered Mapping of population Receptive Fields (GEM-pRF) software implementation. By orthogonalizing the design matrix, our approach enables the direct and fast computation of the objective function’s derivatives, which are used to eliminate the iterative refinement process. This approach dramatically accelerates pRF estimation with high accuracy. Validation using empirical and simulated data confirms GEM-pRF’s accuracy, and benchmarking against established tools demonstrates a reduction in computation time of almost two orders of magnitude. With its modular and extensible design, GEM-pRF provides a critical advancement for large-scale fMRI retinotopic mapping. Furthermore, our reformulated GLM approach in combination with GPU-based implementation offers a broadly applicable solution that may extend beyond visual neuroscience, accelerating computational modelling across various domains in neuroimaging and beyond.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103891"},"PeriodicalIF":11.8,"publicationDate":"2025-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145611717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatial transcriptomics expression prediction from histopathology based on cross-modal mask reconstruction and contrastive learning 基于交叉模态掩模重建和对比学习的组织病理学空间转录组学表达预测
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-25 DOI: 10.1016/j.media.2025.103889
Junzhuo Liu , Markus Eckstein , Zhixiang Wang , Friedrich Feuerhake , Dorit Merhof
Spatial transcriptomics is a technology that captures gene expression at different spatial locations, widely used in tumor microenvironment analysis and molecular profiling of histopathology, providing valuable insights into resolving gene expression and clinical diagnosis of cancer. Due to the high cost of data acquisition, large-scale spatial transcriptomics data remain challenging to obtain. In this study, we develop a contrastive learning-based deep learning method to predict spatially resolved gene expression from the whole-slide images (WSIs). Unlike existing end-to-end prediction frameworks, our method leverages multi-modal contrastive learning to establish a correspondence between histopathological morphology and spatial gene expression in the feature space. By computing cross-modal feature similarity, our method generates spatially resolved gene expression directly from WSIs. Furthermore, to enhance the standard contrastive learning paradigm, a cross-modal masked reconstruction is designed as a pretext task, enabling feature-level fusion between modalities. Notably, our method does not rely on large-scale pretraining datasets or abstract semantic representations from either modality, making it particularly effective for scenarios with limited spatial transcriptomics data. Evaluation across six different disease datasets demonstrates that, compared to existing studies, our method improves Pearson Correlation Coefficient (PCC) in the prediction of highly expressed genes, highly variable genes, and marker genes by 6.27 %, 6.11 %, and 11.26 % respectively. Further analysis indicates that our method preserves gene-gene correlations and applies to datasets with limited samples. Additionally, our method exhibits potential in cancer tissue localization based on biomarker expression. The code repository for this work is available at https://github.com/ngfufdrdh/CMRCNet.
空间转录组学是一种捕获不同空间位置基因表达的技术,广泛应用于肿瘤微环境分析和组织病理学分子谱分析,为解决基因表达和癌症临床诊断提供有价值的见解。由于数据采集成本高,获得大规模空间转录组学数据仍然具有挑战性。在这项研究中,我们开发了一种基于对比学习的深度学习方法,从全幻灯片图像(wsi)中预测空间分辨基因表达。与现有的端到端预测框架不同,我们的方法利用多模态对比学习在特征空间中建立组织病理形态和空间基因表达之间的对应关系。通过计算跨模态特征相似度,我们的方法直接从wsi中生成空间分解的基因表达。此外,为了增强标准的对比学习范式,设计了一个跨模态掩模重构作为借口任务,实现了模态之间的特征级融合。值得注意的是,我们的方法不依赖于大规模的预训练数据集,也不依赖于任何一种模式的抽象语义表示,这使得它在空间转录组学数据有限的情况下特别有效。对6个不同疾病数据集的评估表明,与现有研究相比,我们的方法在预测高表达基因、高可变基因和标记基因方面的Pearson相关系数(PCC)分别提高了6.27%、6.11%和11.26%。进一步的分析表明,我们的方法保留了基因-基因的相关性,并适用于有限样本的数据集。此外,我们的方法显示出基于生物标志物表达的癌症组织定位的潜力。这项工作的代码存储库可从https://github.com/ngfufdrdh/CMRCNet获得。
{"title":"Spatial transcriptomics expression prediction from histopathology based on cross-modal mask reconstruction and contrastive learning","authors":"Junzhuo Liu ,&nbsp;Markus Eckstein ,&nbsp;Zhixiang Wang ,&nbsp;Friedrich Feuerhake ,&nbsp;Dorit Merhof","doi":"10.1016/j.media.2025.103889","DOIUrl":"10.1016/j.media.2025.103889","url":null,"abstract":"<div><div>Spatial transcriptomics is a technology that captures gene expression at different spatial locations, widely used in tumor microenvironment analysis and molecular profiling of histopathology, providing valuable insights into resolving gene expression and clinical diagnosis of cancer. Due to the high cost of data acquisition, large-scale spatial transcriptomics data remain challenging to obtain. In this study, we develop a contrastive learning-based deep learning method to predict spatially resolved gene expression from the whole-slide images (WSIs). Unlike existing end-to-end prediction frameworks, our method leverages multi-modal contrastive learning to establish a correspondence between histopathological morphology and spatial gene expression in the feature space. By computing cross-modal feature similarity, our method generates spatially resolved gene expression directly from WSIs. Furthermore, to enhance the standard contrastive learning paradigm, a cross-modal masked reconstruction is designed as a pretext task, enabling feature-level fusion between modalities. Notably, our method does not rely on large-scale pretraining datasets or abstract semantic representations from either modality, making it particularly effective for scenarios with limited spatial transcriptomics data. Evaluation across six different disease datasets demonstrates that, compared to existing studies, our method improves Pearson Correlation Coefficient (PCC) in the prediction of highly expressed genes, highly variable genes, and marker genes by 6.27 %, 6.11 %, and 11.26 % respectively. Further analysis indicates that our method preserves gene-gene correlations and applies to datasets with limited samples. Additionally, our method exhibits potential in cancer tissue localization based on biomarker expression. The code repository for this work is available at <span><span>https://github.com/ngfufdrdh/CMRCNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"108 ","pages":"Article 103889"},"PeriodicalIF":11.8,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145592833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised learning of spatially varying regularization for diffeomorphic image registration 差分图像配准中空间变化正则化的无监督学习
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-24 DOI: 10.1016/j.media.2025.103887
Junyu Chen , Shuwen Wei , Yihao Liu , Zhangxing Bian , Yufan He , Aaron Carass , Harrison Bai , Yong Du
Spatially varying regularization accommodates the deformation variations that may be necessary for different anatomical regions during deformable image registration. Historically, optimization-based registration models have harnessed spatially varying regularization to address anatomical subtleties. However, most modern deep learning-based models tend to gravitate towards spatially invariant regularization, wherein a homogenous regularization strength is applied across the entire image, potentially disregarding localized variations. In this paper, we propose a hierarchical probabilistic model that integrates a prior distribution on the deformation regularization strength, enabling the end-to-end learning of a spatially varying deformation regularizer directly from the data. The proposed method is straightforward to implement and easily integrates with various registration network architectures. Additionally, automatic tuning of hyperparameters is achieved through Bayesian optimization, allowing efficient identification of optimal hyperparameters for any given registration task. Comprehensive evaluations on publicly available datasets demonstrate that the proposed method significantly improves registration performance and enhances the interpretability of deep learning-based registration, all while maintaining smooth deformations. Our code is freely available at http://bit.ly/3BrXGxz.
空间变化正则化适应在可变形图像配准期间不同解剖区域可能需要的变形变化。从历史上看,基于优化的配准模型利用空间变化的正则化来解决解剖学的微妙之处。然而,大多数现代基于深度学习的模型倾向于空间不变正则化,其中在整个图像上应用均匀的正则化强度,可能忽略局部变化。在本文中,我们提出了一个分层概率模型,该模型集成了变形正则化强度的先验分布,从而可以直接从数据中端到端学习空间变化的变形正则化器。该方法实现简单,易于与各种注册网络体系结构集成。此外,超参数的自动调优是通过贝叶斯优化实现的,允许有效地识别任何给定配准任务的最优超参数。对公开可用数据集的综合评估表明,该方法显著提高了配准性能,增强了基于深度学习的配准的可解释性,同时保持了平滑的形变。我们的代码可以在http://bit.ly/3BrXGxz上免费获得。
{"title":"Unsupervised learning of spatially varying regularization for diffeomorphic image registration","authors":"Junyu Chen ,&nbsp;Shuwen Wei ,&nbsp;Yihao Liu ,&nbsp;Zhangxing Bian ,&nbsp;Yufan He ,&nbsp;Aaron Carass ,&nbsp;Harrison Bai ,&nbsp;Yong Du","doi":"10.1016/j.media.2025.103887","DOIUrl":"10.1016/j.media.2025.103887","url":null,"abstract":"<div><div>Spatially varying regularization accommodates the deformation variations that may be necessary for different anatomical regions during deformable image registration. Historically, optimization-based registration models have harnessed spatially varying regularization to address anatomical subtleties. However, most modern deep learning-based models tend to gravitate towards spatially invariant regularization, wherein a homogenous regularization strength is applied across the entire image, potentially disregarding localized variations. In this paper, we propose a hierarchical probabilistic model that integrates a prior distribution on the deformation regularization strength, enabling the end-to-end learning of a spatially varying deformation regularizer directly from the data. The proposed method is straightforward to implement and easily integrates with various registration network architectures. Additionally, automatic tuning of hyperparameters is achieved through Bayesian optimization, allowing efficient identification of optimal hyperparameters for any given registration task. Comprehensive evaluations on publicly available datasets demonstrate that the proposed method significantly improves registration performance and enhances the interpretability of deep learning-based registration, all while maintaining smooth deformations. Our code is freely available at <span><span>http://bit.ly/3BrXGxz</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"108 ","pages":"Article 103887"},"PeriodicalIF":11.8,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145592834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Medical image analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1