首页 > 最新文献

Medical image analysis最新文献

英文 中文
Channel-Wise Joint Disentanglement Representation Learning for B-Mode and Super-Resolution Ultrasound Based CAD of Breast Cancer 基于b模式和超分辨率超声的乳腺癌CAD的通道联合解缠表示学习
IF 10.9 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-22 DOI: 10.1016/j.media.2026.103957
Yuhang Zheng, Jiale Xu, Qing Hua, Xiaohong Jia, Xueqin Hou, Yanfeng Yao, Zheng Wei, Yulu Zhang, Fanggang Wu, Wei Guo, Yuan Tian, Jun Wang, Shujun Xia, Yijie Dong, Jun Shi, Jianqiao Zhou
{"title":"Channel-Wise Joint Disentanglement Representation Learning for B-Mode and Super-Resolution Ultrasound Based CAD of Breast Cancer","authors":"Yuhang Zheng, Jiale Xu, Qing Hua, Xiaohong Jia, Xueqin Hou, Yanfeng Yao, Zheng Wei, Yulu Zhang, Fanggang Wu, Wei Guo, Yuan Tian, Jun Wang, Shujun Xia, Yijie Dong, Jun Shi, Jianqiao Zhou","doi":"10.1016/j.media.2026.103957","DOIUrl":"https://doi.org/10.1016/j.media.2026.103957","url":null,"abstract":"","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"193 1","pages":""},"PeriodicalIF":10.9,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146032815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anatomy-guided prompting with cross-modal self-alignment for whole-body PET-CT breast cancer segmentation 解剖引导提示与跨模态自对准用于全身PET-CT乳腺癌分割
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-22 DOI: 10.1016/j.media.2026.103956
Jiaju Huang , Xiao Yang , Xinglong Liang , Shaobin Chen , Yue Sun , Greta Sp Mok , Shuo Li , Ying Wang , Tao Tan
Accurate segmentation of breast cancer in PET-CT images is crucial for precise staging, monitoring treatment response, and guiding personalized therapy. However, the small size and dispersed nature of metastatic lesions, coupled with the scarcity of annotated data and heterogeneity between modalities that hinders effective information fusion, make this task challenging. This paper proposes a novel anatomy-guided cross-modal learning framework to address these issues. Our approach first generates organ pseudo-labels through a teacher-student learning paradigm, which serve as anatomical prompts to guide cancer segmentation. We then introduce a self-aligning cross-modal pre-training method that aligns PET and CT features in a shared latent space through masked 3D patch reconstruction, enabling effective cross-modal feature fusion. Finally, we initialize the segmentation network’s encoder with the pre-trained encoder weights, and incorporate organ labels through a Mamba-based prompt encoder and Hypernet-Controlled Cross-Attention mechanism for dynamic anatomical feature extraction and fusion. Notably, our method outperforms eight state-of-the-art methods, including CNN-based, transformer-based, and Mamba-based approaches, on two datasets encompassing primary breast cancer, metastatic breast cancer, and other types of cancer segmentation tasks.
在PET-CT图像中准确分割乳腺癌对于精确分期、监测治疗反应和指导个性化治疗至关重要。然而,转移病灶的小尺寸和分散性质,加上注释数据的缺乏和模式之间的异质性,阻碍了有效的信息融合,使得这项任务具有挑战性。本文提出了一种新的解剖学指导的跨模态学习框架来解决这些问题。我们的方法首先通过师生学习范式生成器官伪标签,作为指导癌症分割的解剖提示。然后,我们引入了一种自对齐的跨模态预训练方法,该方法通过掩膜3D补丁重建在共享潜在空间中对齐PET和CT特征,从而实现有效的跨模态特征融合。最后,我们使用预训练的编码器权重初始化分割网络的编码器,并通过基于mamba的提示编码器和hypernet控制的交叉注意机制合并器官标签,进行动态解剖特征提取和融合。值得注意的是,我们的方法在包括原发性乳腺癌、转移性乳腺癌和其他类型的癌症分割任务的两个数据集上优于八种最先进的方法,包括基于cnn、基于transformer和基于mamba的方法。
{"title":"Anatomy-guided prompting with cross-modal self-alignment for whole-body PET-CT breast cancer segmentation","authors":"Jiaju Huang ,&nbsp;Xiao Yang ,&nbsp;Xinglong Liang ,&nbsp;Shaobin Chen ,&nbsp;Yue Sun ,&nbsp;Greta Sp Mok ,&nbsp;Shuo Li ,&nbsp;Ying Wang ,&nbsp;Tao Tan","doi":"10.1016/j.media.2026.103956","DOIUrl":"10.1016/j.media.2026.103956","url":null,"abstract":"<div><div>Accurate segmentation of breast cancer in PET-CT images is crucial for precise staging, monitoring treatment response, and guiding personalized therapy. However, the small size and dispersed nature of metastatic lesions, coupled with the scarcity of annotated data and heterogeneity between modalities that hinders effective information fusion, make this task challenging. This paper proposes a novel anatomy-guided cross-modal learning framework to address these issues. Our approach first generates organ pseudo-labels through a teacher-student learning paradigm, which serve as anatomical prompts to guide cancer segmentation. We then introduce a self-aligning cross-modal pre-training method that aligns PET and CT features in a shared latent space through masked 3D patch reconstruction, enabling effective cross-modal feature fusion. Finally, we initialize the segmentation network’s encoder with the pre-trained encoder weights, and incorporate organ labels through a Mamba-based prompt encoder and Hypernet-Controlled Cross-Attention mechanism for dynamic anatomical feature extraction and fusion. Notably, our method outperforms eight state-of-the-art methods, including CNN-based, transformer-based, and Mamba-based approaches, on two datasets encompassing primary breast cancer, metastatic breast cancer, and other types of cancer segmentation tasks.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103956"},"PeriodicalIF":11.8,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146033893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UNISELF: A unified network with instance normalization and self-ensembled lesion fusion for multiple sclerosis lesion segmentation UNISELF:一个具有实例归一化和自集合病灶融合的用于多发性硬化症病灶分割的统一网络
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-20 DOI: 10.1016/j.media.2026.103954
Jinwei Zhang , Lianrui Zuo , Blake E. Dewey , Samuel W. Remedios , Yihao Liu , Savannah P. Hays , Dzung L. Pham , Ellen M. Mowry , Scott D. Newsome , Peter A. Calabresi , Shiv Saidha , Aaron Carass , Jerry L. Prince
Automated segmentation of multiple sclerosis (MS) lesions using multicontrast magnetic resonance (MR) images improves efficiency and reproducibility compared to manual delineation, with deep learning (DL) methods achieving state-of-the-art performance. However, these DL-based methods have yet to simultaneously optimize in-domain accuracy and out-of-domain generalization when trained on a single source with limited data, or their performance has been unsatisfactory. To fill this gap, we propose a method called UNISELF, which achieves high accuracy within a single training domain while demonstrating strong generalizability across multiple out-of-domain test datasets. UNISELF employs a novel test-time self-ensembled lesion fusion to improve segmentation accuracy, and leverages test-time instance normalization (TTIN) of latent features to address domain shifts and missing input contrasts. Trained on the ISBI 2015 longitudinal MS segmentation challenge training dataset, UNISELF ranks among the best-performing methods on the challenge test dataset. Additionally, UNISELF outperforms all benchmark methods trained on the same ISBI training data across diverse out-of-domain test datasets with domain shifts and missing contrasts, including the public MICCAI 2016 and UMCL datasets, as well as a private multisite dataset. These test datasets exhibit domain shifts and/or missing contrasts caused by variations in acquisition protocols, scanner types, and imaging artifacts arising from imperfect acquisition. Our code is available at https://github.com/Jinwei1209/UNISELF.
与人工分割相比,使用多层对比磁共振(MR)图像对多发性硬化症(MS)病变进行自动分割提高了效率和可重复性,深度学习(DL)方法实现了最先进的性能。然而,这些基于dl的方法在单一数据源和有限数据上训练时,还没有同时优化域内精度和域外泛化,或者它们的性能令人不满意。为了填补这一空白,我们提出了一种称为UNISELF的方法,该方法在单个训练域中实现了高准确性,同时在多个域外测试数据集上展示了强大的泛化能力。UNISELF采用一种新颖的测试时间自集合病灶融合来提高分割精度,并利用潜在特征的测试时间实例归一化(TTIN)来解决域偏移和缺失输入对比问题。在ISBI 2015纵向MS分割挑战训练数据集上训练,UNISELF是挑战测试数据集上表现最好的方法之一。此外,UNISELF在不同域外测试数据集(包括公共MICCAI 2016和UMCL数据集以及私有多站点数据集)上训练的相同ISBI训练数据优于所有基准方法,这些数据集具有域移位和缺失对比。这些测试数据集表现出由采集协议、扫描仪类型和不完美采集引起的成像工件的变化引起的域移位和/或缺失对比。我们的代码可在https://github.com/Jinwei1209/UNISELF上获得。
{"title":"UNISELF: A unified network with instance normalization and self-ensembled lesion fusion for multiple sclerosis lesion segmentation","authors":"Jinwei Zhang ,&nbsp;Lianrui Zuo ,&nbsp;Blake E. Dewey ,&nbsp;Samuel W. Remedios ,&nbsp;Yihao Liu ,&nbsp;Savannah P. Hays ,&nbsp;Dzung L. Pham ,&nbsp;Ellen M. Mowry ,&nbsp;Scott D. Newsome ,&nbsp;Peter A. Calabresi ,&nbsp;Shiv Saidha ,&nbsp;Aaron Carass ,&nbsp;Jerry L. Prince","doi":"10.1016/j.media.2026.103954","DOIUrl":"10.1016/j.media.2026.103954","url":null,"abstract":"<div><div>Automated segmentation of multiple sclerosis (MS) lesions using multicontrast magnetic resonance (MR) images improves efficiency and reproducibility compared to manual delineation, with deep learning (DL) methods achieving state-of-the-art performance. However, these DL-based methods have yet to simultaneously optimize in-domain accuracy and out-of-domain generalization when trained on a single source with limited data, or their performance has been unsatisfactory. To fill this gap, we propose a method called UNISELF, which achieves high accuracy within a single training domain while demonstrating strong generalizability across multiple out-of-domain test datasets. UNISELF employs a novel test-time self-ensembled lesion fusion to improve segmentation accuracy, and leverages test-time instance normalization (TTIN) of latent features to address domain shifts and missing input contrasts. Trained on the ISBI 2015 longitudinal MS segmentation challenge training dataset, UNISELF ranks among the best-performing methods on the challenge test dataset. Additionally, UNISELF outperforms all benchmark methods trained on the same ISBI training data across diverse out-of-domain test datasets with domain shifts and missing contrasts, including the public MICCAI 2016 and UMCL datasets, as well as a private multisite dataset. These test datasets exhibit domain shifts and/or missing contrasts caused by variations in acquisition protocols, scanner types, and imaging artifacts arising from imperfect acquisition. Our code is available at <span><span>https://github.com/Jinwei1209/UNISELF</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103954"},"PeriodicalIF":11.8,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146014238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VHU-Net: Variational hadamard U-Net for body MRI bias field correction VHU-Net:用于身体MRI偏场校正的变分Hadamard U-Net
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-20 DOI: 10.1016/j.media.2026.103955
Xin Zhu , Ahmet Enis Cetin , Gorkem Durak , Batuhan Gundogdu , Ziliang Hong , Hongyi Pan , Ertugrul Aktas , Elif Keles , Hatice Savas , Aytekin Oto , Hiten Patel , Adam B. Murphy , Ashley Ross , Frank Miller , Baris Turkbey , Ulas Bagci
Bias field artifacts in magnetic resonance imaging (MRI) scans introduce spatially smooth intensity inhomogeneities that degrade image quality and hinder downstream analysis. To address this challenge, we propose a novel variational Hadamard U-Net (VHU-Net) for effective body MRI bias field correction. The encoder comprises multiple convolutional Hadamard transform blocks (ConvHTBlocks), each integrating convolutional layers with a Hadamard transform (HT) layer. Specifically, the HT layer performs channel-wise frequency decomposition to isolate low-frequency components, while a subsequent scaling layer and semi-soft thresholding mechanism suppress redundant high-frequency noise. To compensate for the HT layer’s inability to model inter-channel dependencies, the decoder incorporates an inverse HT-reconstructed transformer block, enabling global, frequency-aware attention for the recovery of spatially consistent bias fields. The stacked decoder ConvHTBlocks further enhance the capacity to reconstruct the underlying ground-truth bias field. Building on the principles of variational inference, we formulate a new evidence lower bound (ELBO) as the training objective, promoting sparsity in the latent space while ensuring accurate bias field estimation. Comprehensive experiments on body MRI datasets demonstrate the superiority of VHU-Net over existing state-of-the-art methods in terms of intensity uniformity. Moreover, the corrected images yield substantial downstream improvements in segmentation accuracy. Our framework offers computational efficiency, interpretability, and robust performance across multi-center datasets, making it suitable for clinical deployment. The codes are available at https://github.com/Holmes696/Probabilistic-Hadamard-U-Net.
磁共振成像(MRI)扫描中的偏置场伪影会引入空间平滑强度不均匀性,从而降低图像质量并阻碍下游分析。为了解决这一挑战,我们提出了一种新的变分Hadamard U-Net (VHU-Net),用于有效的身体MRI偏场校正。编码器包括多个卷积哈达玛变换块(ConvHTBlocks),每个块将卷积层与哈达玛变换(HT)层集成。具体来说,HT层执行按信道的频率分解以隔离低频分量,而随后的缩放层和半软阈值机制抑制冗余的高频噪声。为了弥补高频层无法模拟信道间依赖关系的缺陷,解码器集成了一个逆高频重构的变压器块,实现了对空间一致偏置场恢复的全局频率感知关注。堆叠的解码器ConvHTBlocks进一步增强了重建底层地基真值偏置场的能力。在变分推理原理的基础上,我们制定了一个新的证据下限(ELBO)作为训练目标,在保证准确的偏差场估计的同时,提高了潜在空间的稀疏性。在人体MRI数据集上的综合实验表明,VHU-Net在强度均匀性方面优于现有的最先进方法。此外,校正后的图像在分割精度方面产生了实质性的下游改进。我们的框架提供了跨多中心数据集的计算效率、可解释性和健壮的性能,使其适合临床部署。代码可在https://github.com/Holmes696/Probabilistic-Hadamard-U-Net上获得。
{"title":"VHU-Net: Variational hadamard U-Net for body MRI bias field correction","authors":"Xin Zhu ,&nbsp;Ahmet Enis Cetin ,&nbsp;Gorkem Durak ,&nbsp;Batuhan Gundogdu ,&nbsp;Ziliang Hong ,&nbsp;Hongyi Pan ,&nbsp;Ertugrul Aktas ,&nbsp;Elif Keles ,&nbsp;Hatice Savas ,&nbsp;Aytekin Oto ,&nbsp;Hiten Patel ,&nbsp;Adam B. Murphy ,&nbsp;Ashley Ross ,&nbsp;Frank Miller ,&nbsp;Baris Turkbey ,&nbsp;Ulas Bagci","doi":"10.1016/j.media.2026.103955","DOIUrl":"10.1016/j.media.2026.103955","url":null,"abstract":"<div><div>Bias field artifacts in magnetic resonance imaging (MRI) scans introduce spatially smooth intensity inhomogeneities that degrade image quality and hinder downstream analysis. To address this challenge, we propose a novel variational Hadamard U-Net (VHU-Net) for effective body MRI bias field correction. The encoder comprises multiple convolutional Hadamard transform blocks (ConvHTBlocks), each integrating convolutional layers with a Hadamard transform (HT) layer. Specifically, the HT layer performs channel-wise frequency decomposition to isolate low-frequency components, while a subsequent scaling layer and semi-soft thresholding mechanism suppress redundant high-frequency noise. To compensate for the HT layer’s inability to model inter-channel dependencies, the decoder incorporates an inverse HT-reconstructed transformer block, enabling global, frequency-aware attention for the recovery of spatially consistent bias fields. The stacked decoder ConvHTBlocks further enhance the capacity to reconstruct the underlying ground-truth bias field. Building on the principles of variational inference, we formulate a new evidence lower bound (ELBO) as the training objective, promoting sparsity in the latent space while ensuring accurate bias field estimation. Comprehensive experiments on body MRI datasets demonstrate the superiority of VHU-Net over existing state-of-the-art methods in terms of intensity uniformity. Moreover, the corrected images yield substantial downstream improvements in segmentation accuracy. Our framework offers computational efficiency, interpretability, and robust performance across multi-center datasets, making it suitable for clinical deployment. The codes are available at <span><span>https://github.com/Holmes696/Probabilistic-Hadamard-U-Net</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103955"},"PeriodicalIF":11.8,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146014237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hippocampal surface morphological variation-based genome-wide association analysis network for biomarker detection of Alzheimer’s disease 基于海马表面形态变化的全基因组关联分析网络用于阿尔茨海默病的生物标志物检测
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-18 DOI: 10.1016/j.media.2026.103952
Xiumei Chen , Xinyue Zhang , Wei Xiong , Tao Wang , Aiwei Jia , Qianjin Feng , Meiyan Huang
Performing genome-wide association analysis (GWAS) between hippocampus and whole-genome data can facilitate disease-related biomarker detection of Alzheimer’s disease (AD). However, most existing studies have prioritized hippocampal volume changes and ignored the morphological variations and subfield differences of the hippocampus in AD progression. This disregard restricts the comprehensive understanding of the associations between hippocampus and whole-genome data, which may result in some potentially specific biomarkers of AD being missed. Moreover, the representation of the complex associations between ultra-high-dimensional imaging and whole-genome data remains an unresolved problem in GWAS. To address these issues, we propose an end-to-end hippocampal surface morphological variation-based genome-wide association analysis network (HSM-GWAS) to explore the nonlinear associations between hippocampal surface morphological variations and whole-genome data for AD-related biomarker detection. First, a multi-modality feature extraction module that includes a graph convolution network and an improved diet network is presented to extract imaging and genetic features from non-Euclidean hippocampal surface and whole-genome data, respectively. Second, a dual contrastive learning-based association analysis module is introduced to map and align genetic features to imaging features, thus narrowing the gap between these features and helping explore the complex associations between hippocampal and whole-genome data. Last, a dual cross-attention fusion module is applied to combine imaging and genetic features for disease diagnosis and biomarker detection of AD. Extensive experiments on the real Alzheimer’s Disease Neuroimaging Initiative dataset and simulated data demonstrate that HSM-GWAS considerably improves biomarker detection and disease diagnosis. These findings highlight the ability of HSM-GWAS to discover disease-related biomarkers, suggesting its potential to provide new insights into pathological mechanisms and aid in AD diagnosis. The codes are to be made publicly available at https://github.com/Meiyan88/HSM-GWAS.
在海马和全基因组数据之间进行全基因组关联分析(GWAS)可以促进阿尔茨海默病(AD)疾病相关生物标志物的检测。然而,大多数现有的研究都优先考虑了海马体积的变化,而忽略了阿尔茨海默病进展过程中海马的形态变化和亚区差异。这种忽视限制了对海马体和全基因组数据之间关联的全面理解,这可能导致阿尔茨海默病的一些潜在特异性生物标志物被遗漏。此外,超高维成像和全基因组数据之间复杂关联的表征在GWAS中仍然是一个未解决的问题。为了解决这些问题,我们提出了一个端到端的基于海马表面形态变化的全基因组关联分析网络(HSM-GWAS),以探索海马表面形态变化与ad相关生物标志物检测的全基因组数据之间的非线性关联。首先,提出了包含图卷积网络和改进饮食网络的多模态特征提取模块,分别从非欧几里得海马表面和全基因组数据中提取成像特征和遗传特征。其次,引入了基于双对比学习的关联分析模块,将遗传特征与成像特征进行映射和比对,从而缩小这些特征之间的差距,并有助于探索海马与全基因组数据之间的复杂关联。最后,采用双交叉关注融合模块,结合影像学特征和遗传特征,对AD进行疾病诊断和生物标志物检测。在真实阿尔茨海默病神经成像倡议数据集和模拟数据上进行的大量实验表明,HSM-GWAS大大提高了生物标志物的检测和疾病诊断。这些发现强调了HSM-GWAS发现疾病相关生物标志物的能力,表明其可能为阿尔茨海默病的病理机制提供新的见解,并有助于阿尔茨海默病的诊断。这些代码将在https://github.com/Meiyan88/HSM-GWAS上公开发布。
{"title":"Hippocampal surface morphological variation-based genome-wide association analysis network for biomarker detection of Alzheimer’s disease","authors":"Xiumei Chen ,&nbsp;Xinyue Zhang ,&nbsp;Wei Xiong ,&nbsp;Tao Wang ,&nbsp;Aiwei Jia ,&nbsp;Qianjin Feng ,&nbsp;Meiyan Huang","doi":"10.1016/j.media.2026.103952","DOIUrl":"10.1016/j.media.2026.103952","url":null,"abstract":"<div><div>Performing genome-wide association analysis (GWAS) between hippocampus and whole-genome data can facilitate disease-related biomarker detection of Alzheimer’s disease (AD). However, most existing studies have prioritized hippocampal volume changes and ignored the morphological variations and subfield differences of the hippocampus in AD progression. This disregard restricts the comprehensive understanding of the associations between hippocampus and whole-genome data, which may result in some potentially specific biomarkers of AD being missed. Moreover, the representation of the complex associations between ultra-high-dimensional imaging and whole-genome data remains an unresolved problem in GWAS. To address these issues, we propose an end-to-end hippocampal surface morphological variation-based genome-wide association analysis network (HSM-GWAS) to explore the nonlinear associations between hippocampal surface morphological variations and whole-genome data for AD-related biomarker detection. First, a multi-modality feature extraction module that includes a graph convolution network and an improved diet network is presented to extract imaging and genetic features from non-Euclidean hippocampal surface and whole-genome data, respectively. Second, a dual contrastive learning-based association analysis module is introduced to map and align genetic features to imaging features, thus narrowing the gap between these features and helping explore the complex associations between hippocampal and whole-genome data. Last, a dual cross-attention fusion module is applied to combine imaging and genetic features for disease diagnosis and biomarker detection of AD. Extensive experiments on the real Alzheimer’s Disease Neuroimaging Initiative dataset and simulated data demonstrate that HSM-GWAS considerably improves biomarker detection and disease diagnosis. These findings highlight the ability of HSM-GWAS to discover disease-related biomarkers, suggesting its potential to provide new insights into pathological mechanisms and aid in AD diagnosis. The codes are to be made publicly available at <span><span>https://github.com/Meiyan88/HSM-GWAS</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103952"},"PeriodicalIF":11.8,"publicationDate":"2026-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145995246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transfer learning from 2D natural images to 4D fMRI brain images via geometric mapping 通过几何映射将学习从2D自然图像转移到4D fMRI脑图像
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-17 DOI: 10.1016/j.media.2026.103949
Kai Gao , Lubin Wang , Liang Li , Xiao Chen , Bin Lu , Yu-Wei Wang , Xue-Ying Li , Zi-Han Wang , Hui-Xian Li , Yi-Fan Liao , Li-Ping Cao , Guan-Mao Chen , Jian-Shan Chen , Tao Chen , Tao-Lin Chen , Yan-Rong Chen , Yu-Qi Cheng , Zhao-Song Chu , Shi-Xian Cui , Xi-Long Cui , Dewen Hu
Functional magnetic resonance imaging (fMRI) allows real-time observation of brain activity through blood oxygen level-dependent (BOLD) signals and is extensively used in studies related to sex classification, age estimation, behavioral measurements prediction, and mental disorder diagnosis. However, the application of deep learning techniques to brain fMRI analysis is hindered by the small sample size of fMRI datasets. Transfer learning offers a solution to this problem, but most existing approaches are designed for large-scale 2D natural images. The heterogeneity between 4D fMRI data and 2D natural images makes direct model transfer infeasible. This study proposes a novel geometric mapping-based fMRI transfer learning method that enables transfer learning from 2D natural images to 4D fMRI brain images, bridging the transfer learning gap between fMRI data and natural images. The proposed Multi-scale Multi-domain Feature Aggregation (MMFA) module extracts effective aggregated features and reduces the dimensionality of fMRI data to 3D space. By treating the cerebral cortex as a folded Riemannian manifold in 3D space and mapping it into 2D space using surface geometric mapping, we make the transfer learning from 2D natural images to 4D brain images possible. Moreover, the topological relationships of the cerebral cortex are maintained with our method, and calculations are performed along the Riemannian manifold of the brain, effectively addressing signal interference problems. The experimental results based on the Human Connectome Project (HCP) dataset demonstrate the effectiveness of the proposed method. Our method achieved state-of-the-art performance in sex classification, age estimation, and behavioral measurement prediction tasks. Moreover, we propose a cascaded transfer learning approach for depression diagnosis, and proved its effectiveness on 23 depression datasets. In summary, the proposed fMRI transfer learning method, which accounts for the structural characteristics of the brain, is promising for applying transfer learning from natural images to brain fMRI images, significantly enhancing the performance in various fMRI analysis tasks.
功能磁共振成像(fMRI)可以通过血氧水平依赖(BOLD)信号实时观察大脑活动,并广泛应用于性别分类、年龄估计、行为测量预测和精神障碍诊断等研究。然而,深度学习技术在脑功能磁共振成像分析中的应用受到功能磁共振成像数据集样本量小的阻碍。迁移学习为这个问题提供了一个解决方案,但是大多数现有的方法都是为大规模的二维自然图像设计的。4D fMRI数据与2D自然图像之间的异质性使得直接模型转移不可行。本研究提出了一种新的基于几何映射的fMRI迁移学习方法,实现了从二维自然图像到四维fMRI脑图像的迁移学习,弥合了fMRI数据与自然图像之间的迁移学习差距。提出的多尺度多域特征聚合(MMFA)模块提取有效的聚合特征,并将fMRI数据降维到三维空间。通过将大脑皮层视为三维空间中的折叠黎曼流形,并使用表面几何映射将其映射到二维空间,我们使从二维自然图像到四维大脑图像的迁移学习成为可能。此外,我们的方法保持了大脑皮层的拓扑关系,并沿着大脑的黎曼流形进行计算,有效地解决了信号干扰问题。基于人类连接组计划(HCP)数据集的实验结果证明了该方法的有效性。我们的方法在性别分类、年龄估计和行为测量预测任务中取得了最先进的性能。此外,我们提出了一种用于抑郁症诊断的级联迁移学习方法,并在23个抑郁症数据集上证明了其有效性。综上所述,本文提出的fMRI迁移学习方法考虑了大脑的结构特征,有望将自然图像的迁移学习应用于大脑fMRI图像,显著提高了各种fMRI分析任务的性能。
{"title":"Transfer learning from 2D natural images to 4D fMRI brain images via geometric mapping","authors":"Kai Gao ,&nbsp;Lubin Wang ,&nbsp;Liang Li ,&nbsp;Xiao Chen ,&nbsp;Bin Lu ,&nbsp;Yu-Wei Wang ,&nbsp;Xue-Ying Li ,&nbsp;Zi-Han Wang ,&nbsp;Hui-Xian Li ,&nbsp;Yi-Fan Liao ,&nbsp;Li-Ping Cao ,&nbsp;Guan-Mao Chen ,&nbsp;Jian-Shan Chen ,&nbsp;Tao Chen ,&nbsp;Tao-Lin Chen ,&nbsp;Yan-Rong Chen ,&nbsp;Yu-Qi Cheng ,&nbsp;Zhao-Song Chu ,&nbsp;Shi-Xian Cui ,&nbsp;Xi-Long Cui ,&nbsp;Dewen Hu","doi":"10.1016/j.media.2026.103949","DOIUrl":"10.1016/j.media.2026.103949","url":null,"abstract":"<div><div>Functional magnetic resonance imaging (fMRI) allows real-time observation of brain activity through blood oxygen level-dependent (BOLD) signals and is extensively used in studies related to sex classification, age estimation, behavioral measurements prediction, and mental disorder diagnosis. However, the application of deep learning techniques to brain fMRI analysis is hindered by the small sample size of fMRI datasets. Transfer learning offers a solution to this problem, but most existing approaches are designed for large-scale 2D natural images. The heterogeneity between 4D fMRI data and 2D natural images makes direct model transfer infeasible. This study proposes a novel geometric mapping-based fMRI transfer learning method that enables transfer learning from 2D natural images to 4D fMRI brain images, bridging the transfer learning gap between fMRI data and natural images. The proposed Multi-scale Multi-domain Feature Aggregation (MMFA) module extracts effective aggregated features and reduces the dimensionality of fMRI data to 3D space. By treating the cerebral cortex as a folded Riemannian manifold in 3D space and mapping it into 2D space using surface geometric mapping, we make the transfer learning from 2D natural images to 4D brain images possible. Moreover, the topological relationships of the cerebral cortex are maintained with our method, and calculations are performed along the Riemannian manifold of the brain, effectively addressing signal interference problems. The experimental results based on the Human Connectome Project (HCP) dataset demonstrate the effectiveness of the proposed method. Our method achieved state-of-the-art performance in sex classification, age estimation, and behavioral measurement prediction tasks. Moreover, we propose a cascaded transfer learning approach for depression diagnosis, and proved its effectiveness on 23 depression datasets. In summary, the proposed fMRI transfer learning method, which accounts for the structural characteristics of the brain, is promising for applying transfer learning from natural images to brain fMRI images, significantly enhancing the performance in various fMRI analysis tasks.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103949"},"PeriodicalIF":11.8,"publicationDate":"2026-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145995243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CATERPillar: a flexible framework for generating white matter numerical substrates with incorporated glial cells 卡特彼勒:一个灵活的框架,生成白质数值基质与合并胶质细胞
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-17 DOI: 10.1016/j.media.2026.103946
Jasmine Nguyen-Duc , Malte Brammerloh , Melina Cherchali , Inès De Riedmatten , Jean-Baptiste Pérot , Jonathan Rafael-Patiño , Ileana O. Jelescu
Monte Carlo diffusion simulations in numerical substrates are valuable for exploring the sensitivity and specificity of the diffusion MRI (dMRI) signal to realistic cell microstructure features. A crucial component of such simulations is the use of numerical phantoms that accurately represent the target tissue, which is in this case, cerebral white matter (WM). This study introduces CATERPillar (Computational Axonal Threading Engine for Realistic Proliferation), a novel method that simulates the mechanics of axonal growth using overlapping spheres as elementary units. CATERPillar facilitates parallel axon development while preventing collisions, offering user control over key structural parameters such as cellular density, undulation, beading and myelination. Its uniqueness lies in its ability to generate not only realistic axonal structures but also realistic glial cells, enhancing the biological fidelity of simulations. We showed that our grown substrates feature distributions of key morphological parameters that agree with those from histological studies. The structural realism of the astrocytic components was quantitatively validated using Sholl analysis. Furthermore, the time-dependent diffusion in the extra- and intra-axonal compartments accurately reflected expected characteristics of short-range disorder, as predicted by theoretical models. CATERPillar is open source and can be used to (a) develop new acquisition schemes that sensitise the MRI signal to unique tissue microstructure features, (b) test the accuracy of a broad range of analytical models, and (c) build a set of substrates to train machine learning models on.
在数值基底上进行蒙特卡罗扩散模拟对于探索扩散核磁共振成像(dMRI)信号对真实细胞微观结构特征的敏感性和特异性具有重要意义。这种模拟的一个关键组成部分是使用数字幻象来准确地代表目标组织,在这种情况下,就是脑白质(WM)。本文介绍了一种以重叠球体为基本单元模拟轴突生长机制的新方法——毛毛虫(Computational Axonal Threading Engine for Realistic Proliferation)。卡特彼勒促进平行轴突的发育,同时防止碰撞,为用户提供对关键结构参数的控制,如细胞密度、波动、串珠和髓鞘形成。它的独特之处在于它不仅能够生成真实的轴突结构,而且能够生成真实的胶质细胞,从而提高了模拟的生物保真度。我们发现,我们培养的底物具有与组织学研究一致的关键形态参数分布。星形细胞成分的结构真实性被定量验证使用肖尔分析。此外,轴突外室和轴突内室的时间依赖性扩散准确地反映了理论模型预测的短期紊乱的预期特征。卡特彼勒是开源的,可用于(a)开发新的采集方案,使MRI信号对独特的组织微观结构特征敏感,(b)测试广泛分析模型的准确性,以及(c)构建一组基板来训练机器学习模型。
{"title":"CATERPillar: a flexible framework for generating white matter numerical substrates with incorporated glial cells","authors":"Jasmine Nguyen-Duc ,&nbsp;Malte Brammerloh ,&nbsp;Melina Cherchali ,&nbsp;Inès De Riedmatten ,&nbsp;Jean-Baptiste Pérot ,&nbsp;Jonathan Rafael-Patiño ,&nbsp;Ileana O. Jelescu","doi":"10.1016/j.media.2026.103946","DOIUrl":"10.1016/j.media.2026.103946","url":null,"abstract":"<div><div>Monte Carlo diffusion simulations in numerical substrates are valuable for exploring the sensitivity and specificity of the diffusion MRI (dMRI) signal to realistic cell microstructure features. A crucial component of such simulations is the use of numerical phantoms that accurately represent the target tissue, which is in this case, cerebral white matter (WM). This study introduces CATERPillar (Computational Axonal Threading Engine for Realistic Proliferation), a novel method that simulates the mechanics of axonal growth using overlapping spheres as elementary units. CATERPillar facilitates parallel axon development while preventing collisions, offering user control over key structural parameters such as cellular density, undulation, beading and myelination. Its uniqueness lies in its ability to generate not only realistic axonal structures but also realistic glial cells, enhancing the biological fidelity of simulations. We showed that our grown substrates feature distributions of key morphological parameters that agree with those from histological studies. The structural realism of the astrocytic components was quantitatively validated using Sholl analysis. Furthermore, the time-dependent diffusion in the extra- and intra-axonal compartments accurately reflected expected characteristics of short-range disorder, as predicted by theoretical models. CATERPillar is open source and can be used to (a) develop new acquisition schemes that sensitise the MRI signal to unique tissue microstructure features, (b) test the accuracy of a broad range of analytical models, and (c) build a set of substrates to train machine learning models on.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"110 ","pages":"Article 103946"},"PeriodicalIF":11.8,"publicationDate":"2026-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145995244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
U2AD: Uncertainty-based unsupervised anomaly detection framework for detecting T2 hyperintensity in MRI spinal cord U2AD:基于不确定性的非监督异常检测框架检测MRI脊髓T2高信号
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-17 DOI: 10.1016/j.media.2026.103939
Qi Zhang , Xiuyuan Chen , Ziyi He , Kun Wang , Lianming Wu , Hongxing Shen , Jianqi Sun
T2 hyperintensities in spinal cord MR images are crucial biomarkers for conditions such as degenerative cervical myelopathy (DCM). However, current clinical diagnoses primarily rely on manual evaluation. Deep learning methods have shown promise in lesion detection, but most supervised approaches are heavily dependent on large, annotated datasets. Unsupervised anomaly detection (UAD) offers a compelling alternative by eliminating the need for abnormal data annotations. However, existing UAD methods face challenges of domain shifts and task conflict. We propose an Uncertainty-based Unsupervised Anomaly Detection framework, termed U2AD, to address these limitations. Unlike traditional methods, U2AD is designed to be trained and tested within the same clinical dataset, following a “mask-and-reconstruction” paradigm built on a Vision Transformer-based architecture. We introduce an uncertainty-guided masking strategy to resolve task conflicts between normal reconstruction and anomaly detection to achieve an optimal balance. Specifically, we employ a Monte-Carlo inference technique to estimate reconstruction uncertainty mappings during training. By iteratively optimizing reconstruction training under the guidance of both epistemic and aleatoric uncertainty, U2AD improves the normal representation learning while maintaining the sensitivity to anomalies. Experimental results demonstrate that U2AD outperforms existing UAD methods in patient-level identification and segment-level localization of spinal cord T2 hyperintensities. This framework establishes a new benchmark for incorporating uncertainty guidance into UAD. Our code is available at: https://github.com/zhibaishouheilab/U2AD
脊髓MR图像中的T2高信号是退行性颈椎病(DCM)等疾病的重要生物标志物。然而,目前的临床诊断主要依赖于人工评估。深度学习方法在病变检测方面已经显示出前景,但大多数监督方法严重依赖于大型、带注释的数据集。无监督异常检测(UAD)通过消除对异常数据注释的需要,提供了一种引人注目的替代方案。然而,现有的UAD方法面临着领域转移和任务冲突的挑战。我们提出了一个基于不确定性的无监督异常检测框架,称为U2AD,以解决这些限制。与传统方法不同,U2AD被设计为在相同的临床数据集中进行训练和测试,遵循基于Vision transformer架构的“掩模和重建”范式。我们引入了一种不确定性引导掩蔽策略来解决正常重建和异常检测之间的任务冲突,以达到最佳平衡。具体来说,我们采用蒙特卡罗推理技术来估计训练过程中的重建不确定性映射。U2AD通过在认知不确定性和任意不确定性的指导下迭代优化重构训练,在保持异常敏感性的同时提高了正常表示学习。实验结果表明,U2AD在患者水平识别和脊髓T2高信号节段水平定位方面优于现有的UAD方法。该框架为将不确定性指导纳入UAD建立了新的基准。我们的代码可在:https://github.com/zhibaishouheilab/U2AD
{"title":"U2AD: Uncertainty-based unsupervised anomaly detection framework for detecting T2 hyperintensity in MRI spinal cord","authors":"Qi Zhang ,&nbsp;Xiuyuan Chen ,&nbsp;Ziyi He ,&nbsp;Kun Wang ,&nbsp;Lianming Wu ,&nbsp;Hongxing Shen ,&nbsp;Jianqi Sun","doi":"10.1016/j.media.2026.103939","DOIUrl":"10.1016/j.media.2026.103939","url":null,"abstract":"<div><div>T2 hyperintensities in spinal cord MR images are crucial biomarkers for conditions such as degenerative cervical myelopathy (DCM). However, current clinical diagnoses primarily rely on manual evaluation. Deep learning methods have shown promise in lesion detection, but most supervised approaches are heavily dependent on large, annotated datasets. Unsupervised anomaly detection (UAD) offers a compelling alternative by eliminating the need for abnormal data annotations. However, existing UAD methods face challenges of domain shifts and task conflict. We propose an <strong>U</strong>ncertainty-based <strong>U</strong>nsupervised <strong>A</strong>nomaly <strong>D</strong>etection framework, termed <em>U</em><sup>2</sup><em>AD</em>, to address these limitations. Unlike traditional methods, <em>U</em><sup>2</sup><em>AD</em> is designed to be trained and tested within the same clinical dataset, following a “mask-and-reconstruction” paradigm built on a Vision Transformer-based architecture. We introduce an uncertainty-guided masking strategy to resolve task conflicts between normal reconstruction and anomaly detection to achieve an optimal balance. Specifically, we employ a Monte-Carlo inference technique to estimate reconstruction uncertainty mappings during training. By iteratively optimizing reconstruction training under the guidance of both epistemic and aleatoric uncertainty, <em>U</em><sup>2</sup><em>AD</em> improves the normal representation learning while maintaining the sensitivity to anomalies. Experimental results demonstrate that <em>U</em><sup>2</sup><em>AD</em> outperforms existing UAD methods in patient-level identification and segment-level localization of spinal cord T2 hyperintensities. This framework establishes a new benchmark for incorporating uncertainty guidance into UAD. Our code is available at: <span><span>https://github.com/zhibaishouheilab/U2AD</span><svg><path></path></svg></span></div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103939"},"PeriodicalIF":11.8,"publicationDate":"2026-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145995245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advances in automated fetal brain MRI segmentation and biometry: Insights from the FeTA 2024 challenge 自动化胎儿脑MRI分割和生物计量的进展:来自FeTA 2024挑战的见解
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-16 DOI: 10.1016/j.media.2026.103941
Vladyslav Zalevskyi , Thomas Sanchez , Misha Kaandorp , Margaux Roulet , Diego Fajardo-Rojas , Liu Li , Jana Hutter , Hongwei Bran Li , Matthew J. Barkovich , Hui Ji , Luca Wilhelmi , Aline Dändliker , Céline Steger , Mériam Koob , Yvan Gomez , Anton Jakovčić , Melita Klaić , Ana Adžić , Pavel Marković , Gracia Grabarić , Meritxell Bach Cuadra
Accurate fetal brain tissue segmentation and biometric measurement are essential for monitoring neurodevelopment and detecting abnormalities in utero. The Fetal Tissue Annotation (FeTA) Challenges have established robust multi-center benchmarks for evaluating state-of-the-art segmentation methods. This paper presents the results of the 2024 challenge edition, which introduced three key innovations.
First, we introduced a topology-aware metric based on the Euler characteristic difference (ED) to overcome the performance plateau observed with traditional metrics like Dice or Hausdorff distance (HD), as the performance of the best models in segmentation surpassed the inter-rater variability. While the best teams reached similar scores in Dice (0.81-0.82) and HD95 (2.1–2.3 mm), ED provided greater discriminative power: the winning method achieved an ED of 20.9, representing roughly a 50% improvement over the second- and third-ranked teams despite comparable Dice scores.
Second, we introduced a new 0.55T low-field MRI test set, which, when paired with high-quality super-resolution reconstruction, achieved the highest segmentation performance across all test cohorts (Dice=0.86, HD95=1.69, ED=6.26). This provides the first quantitative evidence that low-cost, low-field MRI can match or surpass high-field systems in automated fetal brain segmentation.
Third, the new biometry estimation task exposed a clear performance gap: although the best model reached a mean average percentage error (MAPE) of 7.72%, most submissions failed to outperform a simple gestational-age-based linear regression model (MAPE=9.56%), and all remained above inter-rater variability with a MAPE of 5.38%.
Finally, by analyzing the top-performing models from FeTA 2024 alongside those from previous challenge editions, we identify ensembles of 3D nnU-Net trained on both real and synthetic data with both image- and anatomy-level augmentations as the most effective approaches for fetal brain segmentation. Our quantitative analysis reveals that acquisition site, super-resolution strategy, and image quality are the primary sources of domain shift, informing recommendations to enhance the robustness and generalizability of automated fetal brain analysis methods.
准确的胎儿脑组织分割和生物测量对于监测神经发育和检测子宫内异常是必不可少的。胎儿组织注释(FeTA)挑战为评估最先进的分割方法建立了强大的多中心基准。本文介绍了2024挑战版的结果,其中介绍了三个关键创新。首先,我们引入了基于欧拉特征差(ED)的拓扑感知度量,以克服传统度量(如Dice或Hausdorff distance, HD)所观察到的性能平台,因为最佳模型在分割中的性能超过了速率间的可变性。虽然最好的团队在Dice(0.81-0.82)和HD95(2.1-2.3 mm)中获得了相似的分数,但ED提供了更大的辨别能力:获胜方法的ED达到了20.9,比排名第二和第三的团队提高了大约50%,尽管Dice分数相当。其次,我们引入了一个新的0.55T低场MRI测试集,当它与高质量的超分辨率重建配对时,在所有测试队列中获得了最高的分割性能(Dice=0.86, HD95=1.69, ED=6.26)。这提供了第一个定量证据,低成本,低场MRI可以匹配或超过高场系统在胎儿脑自动分割。第三,新的生物特征估计任务暴露出明显的性能差距:尽管最佳模型达到7.72%的平均百分比误差(MAPE),但大多数提交的模型未能优于简单的基于胎龄的线性回归模型(MAPE=9.56%),并且都保持在5.38%的等级间变异之上。最后,通过分析FeTA 2024和以前的挑战版本中表现最好的模型,我们确定了在真实和合成数据上训练的3D nnU-Net集合,并在图像和解剖水平上增强,作为胎儿大脑分割的最有效方法。我们的定量分析表明,采集地点、超分辨率策略和图像质量是域漂移的主要来源,为提高自动化胎儿脑分析方法的鲁棒性和通用性提供了建议。
{"title":"Advances in automated fetal brain MRI segmentation and biometry: Insights from the FeTA 2024 challenge","authors":"Vladyslav Zalevskyi ,&nbsp;Thomas Sanchez ,&nbsp;Misha Kaandorp ,&nbsp;Margaux Roulet ,&nbsp;Diego Fajardo-Rojas ,&nbsp;Liu Li ,&nbsp;Jana Hutter ,&nbsp;Hongwei Bran Li ,&nbsp;Matthew J. Barkovich ,&nbsp;Hui Ji ,&nbsp;Luca Wilhelmi ,&nbsp;Aline Dändliker ,&nbsp;Céline Steger ,&nbsp;Mériam Koob ,&nbsp;Yvan Gomez ,&nbsp;Anton Jakovčić ,&nbsp;Melita Klaić ,&nbsp;Ana Adžić ,&nbsp;Pavel Marković ,&nbsp;Gracia Grabarić ,&nbsp;Meritxell Bach Cuadra","doi":"10.1016/j.media.2026.103941","DOIUrl":"10.1016/j.media.2026.103941","url":null,"abstract":"<div><div>Accurate fetal brain tissue segmentation and biometric measurement are essential for monitoring neurodevelopment and detecting abnormalities in utero. The Fetal Tissue Annotation (FeTA) Challenges have established robust multi-center benchmarks for evaluating state-of-the-art segmentation methods. This paper presents the results of the 2024 challenge edition, which introduced three key innovations.</div><div>First, we introduced a topology-aware metric based on the Euler characteristic difference (ED) to overcome the performance plateau observed with traditional metrics like Dice or Hausdorff distance (HD), as the performance of the best models in segmentation surpassed the inter-rater variability. While the best teams reached similar scores in Dice (0.81-0.82) and HD95 (2.1–2.3 mm), ED provided greater discriminative power: the winning method achieved an ED of 20.9, representing roughly a 50% improvement over the second- and third-ranked teams despite comparable Dice scores.</div><div>Second, we introduced a new 0.55T low-field MRI test set, which, when paired with high-quality super-resolution reconstruction, achieved the highest segmentation performance across all test cohorts (Dice=0.86, HD95=1.69, ED=6.26). This provides the first quantitative evidence that low-cost, low-field MRI can match or surpass high-field systems in automated fetal brain segmentation.</div><div>Third, the new biometry estimation task exposed a clear performance gap: although the best model reached a mean average percentage error (MAPE) of 7.72%, most submissions failed to outperform a simple gestational-age-based linear regression model (MAPE=9.56%), and all remained above inter-rater variability with a MAPE of 5.38%.</div><div>Finally, by analyzing the top-performing models from FeTA 2024 alongside those from previous challenge editions, we identify ensembles of 3D nnU-Net trained on both real and synthetic data with both image- and anatomy-level augmentations as the most effective approaches for fetal brain segmentation. Our quantitative analysis reveals that acquisition site, super-resolution strategy, and image quality are the primary sources of domain shift, informing recommendations to enhance the robustness and generalizability of automated fetal brain analysis methods.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103941"},"PeriodicalIF":11.8,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145995249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AEM: An interpretable multi-task multi-modal framework for cardiac disease prediction AEM:一个可解释的多任务多模式心脏病预测框架
IF 11.8 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-16 DOI: 10.1016/j.media.2026.103951
Jiachuan Peng , Marcel Beetz , Abhirup Banerjee , Min Chen , Vicente Grau
Cardiovascular disease (CVD) is one of the leading causes of death and illness across the world. Especially, early prediction of heart failure (HF) is complicated due to the heterogeneity of its clinical presentations and symptoms. These challenges underscore the need for a multidisciplinary approach for comprehensive evaluation of cardiac state. To this end, we specifically select electrocardiogram (ECG) and 3D cardiac anatomy for their complementary coverage of cardiac electrical activities and fine-grained structural modeling. Building upon this, we present a novel pre-training framework, named Anatomy-Electrocardiogram Model (AEM), to explore their complex interactions. AEM adopts a multi-task self-supervised scheme that combines a masked reconstruction objective with a cardiac measurement (CM) regression branch to embed cardiac functional priors and structural details. Unlike image-domain models that typically localize the whole heart within the image, our 3D anatomy is background-free and continuous in 3D space. Hence, the model can naturally concentrate on finer structures at the patch level. The further integration with ECG captures functional dynamics through electrical conduction, encapsulating holistic cardiac representations. Extensive experiments are conducted on the multi-modal datasets collected from the UK Biobank, which contain paired biventricular point cloud anatomy and 12-lead ECG data. Our proposed AEM achieves an area under the receiver operating characteristic curve of 0.8192 for incident HF prediction and a concordance index of 0.6976 for survival prediction under linear evaluation, outperforming the state-of-the-art multi-modal methods. Additionally, we study the interpretability of the disease prediction by observing that our model effectively recognizes clinically plausible patterns and exhibits a high association with clinical features.
心血管疾病(CVD)是世界上导致死亡和疾病的主要原因之一。尤其是心力衰竭(HF)的早期预测由于其临床表现和症状的异质性而变得复杂。这些挑战强调需要一个多学科的方法来全面评估心脏状态。为此,我们特别选择了心电图(ECG)和3D心脏解剖,因为它们对心脏电活动和细粒度结构建模的补充覆盖。在此基础上,我们提出了一个新的预训练框架,称为解剖-心电图模型(AEM),以探索它们之间复杂的相互作用。AEM采用一种多任务自监督方案,该方案将隐藏重建目标与心脏测量(CM)回归分支相结合,嵌入心脏功能先验和结构细节。与通常在图像中定位整个心脏的图像域模型不同,我们的3D解剖结构在3D空间中是无背景和连续的。因此,该模型可以自然地集中在补丁级别的更精细的结构上。与ECG的进一步整合通过电传导捕获功能动态,封装整体心脏表征。对从英国生物银行收集的多模态数据集进行了广泛的实验,其中包含成对的双心室点云解剖和12导联心电图数据。我们提出的AEM在线性评估下,对事件HF预测的接受者工作特征曲线下面积为0.8192,对生存预测的一致性指数为0.6976,优于目前最先进的多模态方法。此外,我们研究了疾病预测的可解释性,观察到我们的模型有效地识别临床合理的模式,并表现出与临床特征的高度关联。
{"title":"AEM: An interpretable multi-task multi-modal framework for cardiac disease prediction","authors":"Jiachuan Peng ,&nbsp;Marcel Beetz ,&nbsp;Abhirup Banerjee ,&nbsp;Min Chen ,&nbsp;Vicente Grau","doi":"10.1016/j.media.2026.103951","DOIUrl":"10.1016/j.media.2026.103951","url":null,"abstract":"<div><div>Cardiovascular disease (CVD) is one of the leading causes of death and illness across the world. Especially, early prediction of heart failure (HF) is complicated due to the heterogeneity of its clinical presentations and symptoms. These challenges underscore the need for a multidisciplinary approach for comprehensive evaluation of cardiac state. To this end, we specifically select electrocardiogram (ECG) and 3D cardiac anatomy for their complementary coverage of cardiac electrical activities and fine-grained structural modeling. Building upon this, we present a novel pre-training framework, named Anatomy-Electrocardiogram Model (AEM), to explore their complex interactions. AEM adopts a multi-task self-supervised scheme that combines a masked reconstruction objective with a cardiac measurement (CM) regression branch to embed cardiac functional priors and structural details. Unlike image-domain models that typically localize the whole heart within the image, our 3D anatomy is background-free and continuous in 3D space. Hence, the model can naturally concentrate on finer structures at the patch level. The further integration with ECG captures functional dynamics through electrical conduction, encapsulating holistic cardiac representations. Extensive experiments are conducted on the multi-modal datasets collected from the UK Biobank, which contain paired biventricular point cloud anatomy and 12-lead ECG data. Our proposed AEM achieves an area under the receiver operating characteristic curve of 0.8192 for incident HF prediction and a concordance index of 0.6976 for survival prediction under linear evaluation, outperforming the state-of-the-art multi-modal methods. Additionally, we study the interpretability of the disease prediction by observing that our model effectively recognizes clinically plausible patterns and exhibits a high association with clinical features.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"109 ","pages":"Article 103951"},"PeriodicalIF":11.8,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145995248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Medical image analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1