Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention最新文献
Pub Date : 2026-01-01Epub Date: 2025-09-21DOI: 10.1007/978-3-032-04947-6_65
Mengqi Wu, Minhui Yu, Weili Lin, Pew-Thian Yap, Mingxia Liu
Multi-site brain MRI heterogeneity caused by differences in scanner field strengths, acquisition protocols, and software versions poses a significant challenge for consistent analysis. Image-level harmonization, leveraging advanced learning methods, has attracted increasing attention. However, existing methods often rely on paired data (e.g., human traveling phantoms) for training, which are not always available. Some methods perform MRI harmonization by transferring target-style features to source images but require explicitly learning disentangled image styles (e.g., contrast) via encoder-decoder networks, which increases computational complexity. This paper presents an unpaired MRI harmonization (UMH) framework based on a new image style-guided diffusion model. UMH operates in two stages: (1) a coarse harmonizer that aligns multi-site MRIs to a unified domain via a conditional latent diffusion model while preserving anatomical content; and (2) a fine harmonizer that adapts coarsely harmonized images to a specific target using style embeddings derived from a pre-trained Contrastive Language-Image Pre-training (CLIP) encoder, which captures semantic style differences between the original MRIs and their coarsely-aligned counterparts, eliminating the need for paired data. By leveraging rich semantic style representations of CLIP, UMH avoids learning image styles explicitly, thereby reducing computation costs. We evaluate UMH on 4,123 MRIs from three distinct multi-site datasets, with results suggesting its superiority over several state-of-the-art (SOTA) methods across image-level comparison, downstream classification, and brain tissue segmentation tasks.
{"title":"Unpaired Multi-Site Brain MRI Harmonization with Image Style-Guided Latent Diffusion.","authors":"Mengqi Wu, Minhui Yu, Weili Lin, Pew-Thian Yap, Mingxia Liu","doi":"10.1007/978-3-032-04947-6_65","DOIUrl":"10.1007/978-3-032-04947-6_65","url":null,"abstract":"<p><p>Multi-site brain MRI heterogeneity caused by differences in scanner field strengths, acquisition protocols, and software versions poses a significant challenge for consistent analysis. Image-level harmonization, leveraging advanced learning methods, has attracted increasing attention. However, existing methods often rely on paired data (<i>e.g.</i>, human traveling phantoms) for training, which are not always available. Some methods perform MRI harmonization by transferring target-style features to source images but require explicitly learning disentangled image styles (<i>e.g.</i>, contrast) via encoder-decoder networks, which increases computational complexity. This paper presents an unpaired MRI harmonization (UMH) framework based on a new image style-guided diffusion model. UMH operates in two stages: (1) a <i>coarse harmonizer</i> that aligns multi-site MRIs to a unified domain via a conditional latent diffusion model while preserving anatomical content; and (2) a <i>fine harmonizer</i> that adapts coarsely harmonized images to a specific target using style embeddings derived from a pre-trained Contrastive Language-Image Pre-training (CLIP) encoder, which captures semantic style differences between the original MRIs and their coarsely-aligned counterparts, eliminating the need for paired data. By leveraging rich semantic style representations of CLIP, UMH avoids learning image styles explicitly, thereby reducing computation costs. We evaluate UMH on 4,123 MRIs from three distinct multi-site datasets, with results suggesting its superiority over several state-of-the-art (SOTA) methods across image-level comparison, downstream classification, and brain tissue segmentation tasks.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15962 ","pages":"683-693"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12706746/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145776983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-01Epub Date: 2025-09-19DOI: 10.1007/978-3-032-05162-2_23
Minheng Chen, Xiaowei Yu, Jing Zhang, Tong Chen, Chao Cao, Yan Zhuang, Yanjun Lyu, Lu Zhang, Tianming Liu, Dajiang Zhu
Understanding the organization of human brain networks has become a central focus in neuroscience, particularly in the study of functional connectivity, which plays a crucial role in diagnosing neurological disorders. Advances in functional magnetic resonance imaging and machine learning techniques have significantly improved brain network analysis. However, traditional machine learning approaches struggle to capture the complex relationships between brain regions, while deep learning methods, particularly Transformer-based models, face computational challenges due to their quadratic complexity in long-sequence modeling. To address these limitations, we propose a Core-Periphery State-Space Model (CP-SSM), an innovative framework for functional connectome classification. Specifically, we introduce Mamba, a selective state-space model with linear complexity, to effectively capture long-range dependencies in functional brain networks. Furthermore, inspired by the core-periphery (CP) organization, a fundamental characteristic of brain networks that enhances efficient information transmission, we design CP-MoE, a CP-guided Mixture-of-Experts that improves the representation learning of brain connectivity patterns. We evaluate CP-SSM on two benchmark fMRI datasets: ABIDE and ADNI. Experimental results demonstrate that CP-SSM surpasses Transformer-based models in classification performance while significantly reducing computational complexity. These findings highlight the effectiveness and efficiency of CP-SSM in modeling brain functional connectivity, offering a promising direction for neuroimaging-based neurological disease diagnosis. Our code is available at https://github.com/m1nhengChen/cpssm.
{"title":"Core-Periphery Principle Guided State Space Model for Functional Connectome Classification.","authors":"Minheng Chen, Xiaowei Yu, Jing Zhang, Tong Chen, Chao Cao, Yan Zhuang, Yanjun Lyu, Lu Zhang, Tianming Liu, Dajiang Zhu","doi":"10.1007/978-3-032-05162-2_23","DOIUrl":"10.1007/978-3-032-05162-2_23","url":null,"abstract":"<p><p>Understanding the organization of human brain networks has become a central focus in neuroscience, particularly in the study of functional connectivity, which plays a crucial role in diagnosing neurological disorders. Advances in functional magnetic resonance imaging and machine learning techniques have significantly improved brain network analysis. However, traditional machine learning approaches struggle to capture the complex relationships between brain regions, while deep learning methods, particularly Transformer-based models, face computational challenges due to their quadratic complexity in long-sequence modeling. To address these limitations, we propose a Core-Periphery State-Space Model (CP-SSM), an innovative framework for functional connectome classification. Specifically, we introduce Mamba, a selective state-space model with linear complexity, to effectively capture long-range dependencies in functional brain networks. Furthermore, inspired by the core-periphery (CP) organization, a fundamental characteristic of brain networks that enhances efficient information transmission, we design CP-MoE, a CP-guided Mixture-of-Experts that improves the representation learning of brain connectivity patterns. We evaluate CP-SSM on two benchmark fMRI datasets: ABIDE and ADNI. Experimental results demonstrate that CP-SSM surpasses Transformer-based models in classification performance while significantly reducing computational complexity. These findings highlight the effectiveness and efficiency of CP-SSM in modeling brain functional connectivity, offering a promising direction for neuroimaging-based neurological disease diagnosis. Our code is available at https://github.com/m1nhengChen/cpssm.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15971 ","pages":"236-246"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12715851/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145807067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-01Epub Date: 2025-09-18DOI: 10.1007/978-3-032-05182-0_31
Zheyu Wen, George Biros
Positron Emission Tomography (PET) is often used to manage Alzheimer's disease (AD). To better understand progression, we introduce and evaluate a mathematical model that couples at parcellated gray matter regions. We term this model LNODE for "latent network ordinary differential equations". At each region, we track normal , abnormal , and latent states that intend to capture unobservable mechanisms coupled to progression. LNODE is parameterized by subject-specific parameters and cohort parameters. We jointly invert for these parameters by fitting the model to -PET data from 585 subjects from the ADNI dataset. Although underparameterized, our model achieves population compared to when fitting without latent states. Furthermore, these preliminary results suggest the existence of different subtypes of progression.
A β正电子发射断层扫描(PET)通常用于治疗阿尔茨海默病(AD)。为了更好地理解A β的进展,我们引入并评估了一个数学模型,该模型将A β偶联在包裹状灰质区域。我们称这个模型为“潜在网络常微分方程”的LNODE。在每个区域,我们跟踪正常A β,异常A β和m潜伏状态,意图捕捉与A β进展相关的不可观察机制。LNODE由特定主题参数和队列参数参数化。通过将模型拟合到来自ADNI数据集的585名受试者的A β -PET数据,我们共同反演了这些参数。虽然我们的模型是欠参数化的,但与没有潜在状态拟合时的r2≤60%相比,我们的模型达到了总体r2≥98%。此外,这些初步结果表明存在不同亚型的A β进展。
{"title":"<ArticleTitle xmlns:ns0=\"http://www.w3.org/1998/Math/MathML\">LNODE: Uncovering the Latent Dynamics of <ns0:math><ns0:mi>A</ns0:mi> <ns0:mi>β</ns0:mi></ns0:math> in Alzheimer's Disease.","authors":"Zheyu Wen, George Biros","doi":"10.1007/978-3-032-05182-0_31","DOIUrl":"10.1007/978-3-032-05182-0_31","url":null,"abstract":"<p><p><math><mtext>A</mtext> <mi>β</mi></math> Positron Emission Tomography (PET) is often used to manage Alzheimer's disease (AD). To better understand <math><mtext>A</mtext> <mi>β</mi></math> progression, we introduce and evaluate a mathematical model that couples <math><mtext>A</mtext> <mi>β</mi></math> at parcellated gray matter regions. We term this model LNODE for \"<i>latent network ordinary differential equations</i>\". At each region, we track normal <math><mtext>A</mtext> <mi>β</mi></math> , abnormal <math><mtext>A</mtext> <mi>β</mi></math> , and <math><mi>m</mi></math> latent states that intend to capture unobservable mechanisms coupled to <math><mtext>A</mtext> <mi>β</mi></math> progression. LNODE is parameterized by subject-specific parameters and cohort parameters. We jointly invert for these parameters by fitting the model to <math><mtext>A</mtext> <mi>β</mi></math> -PET data from 585 subjects from the ADNI dataset. Although underparameterized, our model achieves population <math> <msup><mrow><mi>R</mi></mrow> <mrow><mn>2</mn></mrow> </msup> <mo>≥</mo> <mn>98</mn> <mo>%</mo></math> compared to <math> <msup><mrow><mi>R</mi></mrow> <mrow><mn>2</mn></mrow> </msup> <mo>≤</mo> <mn>60</mn> <mo>%</mo></math> when fitting without latent states. Furthermore, these preliminary results suggest the existence of different subtypes of <math><mtext>A</mtext> <mi>β</mi></math> progression.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15974 ","pages":"313-322"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12784421/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145954748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-01Epub Date: 2025-09-20DOI: 10.1007/978-3-032-04981-0_24
Benjamin D Killeen, Liam J Wang, Blanca Iñígo, Han Zhang, Mehran Armand, Russell H Taylor, Greg Osgood, Mathias Unberath
Language promptable X-ray image segmentation would enable greater flexibility for human-in-the-loop workflows in diagnostic and interventional precision medicine. Prior efforts have contributed task-specific models capable of solving problems within a narrow scope, but expanding to broader use requires additional data, annotations, and training time. Recently, language-aligned foundation models (LFMs) - machine learning models trained on large amounts of highly variable image and text data thus enabling broad applicability - have emerged as promising tools for automated image analysis. Existing foundation models for medical image analysis focus on scenarios and modalities where large, richly annotated datasets are available. However, the X-ray imaging modality features highly variable image appearance and applications, from diagnostic chest X-rays to interventional fluoroscopy, with varying availability of data. To pave the way toward an LFM for comprehensive and language-aligned analysis of arbitrary medical X-ray images, we introduce FluoroSAM, a language-promptable variant of the Segment-Anything Model, trained from scratch on 3M synthetic X-ray images from a wide variety of human anatomies, imaging geometries, and viewing angles. These include pseudo-ground truth masks for 128 organ types and 464 tools with associated text descriptions. FluoroSAM is capable of segmenting myriad anatomical structures and tools based on natural language prompts, thanks to the novel incorporation of vector quantization (VQ) of text embeddings in the training process. We demonstrate FluoroSAM's performance quantitatively on real X-ray images and showcase on several applications how FluoroSAM is a key enabler for rich human-machine interaction in the X-ray image acquisition and analysis context. Code is available at https://github.com/arcadelab/fluorosam.
{"title":"FluoroSAM: A Language-promptable Foundation Model for Flexible X-ray Image Segmentation.","authors":"Benjamin D Killeen, Liam J Wang, Blanca Iñígo, Han Zhang, Mehran Armand, Russell H Taylor, Greg Osgood, Mathias Unberath","doi":"10.1007/978-3-032-04981-0_24","DOIUrl":"10.1007/978-3-032-04981-0_24","url":null,"abstract":"<p><p>Language promptable X-ray image segmentation would enable greater flexibility for human-in-the-loop workflows in diagnostic and interventional precision medicine. Prior efforts have contributed task-specific models capable of solving problems within a narrow scope, but expanding to broader use requires additional data, annotations, and training time. Recently, language-aligned foundation models (LFMs) - machine learning models trained on large amounts of highly variable image and text data thus enabling broad applicability - have emerged as promising tools for automated image analysis. Existing foundation models for medical image analysis focus on scenarios and modalities where large, richly annotated datasets are available. However, the X-ray imaging modality features highly variable image appearance and applications, from diagnostic chest X-rays to interventional fluoroscopy, with varying availability of data. To pave the way toward an LFM for comprehensive and language-aligned analysis of arbitrary medical X-ray images, we introduce FluoroSAM, a language-promptable variant of the Segment-Anything Model, trained from scratch on 3M synthetic X-ray images from a wide variety of human anatomies, imaging geometries, and viewing angles. These include pseudo-ground truth masks for 128 organ types and 464 tools with associated text descriptions. FluoroSAM is capable of segmenting myriad anatomical structures and tools based on natural language prompts, thanks to the novel incorporation of vector quantization (VQ) of text embeddings in the training process. We demonstrate FluoroSAM's performance quantitatively on real X-ray images and showcase on several applications how FluoroSAM is a key enabler for rich human-machine interaction in the X-ray image acquisition and analysis context. <i>Code is available at</i> https://github.com/arcadelab/fluorosam.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15966 ","pages":"248-258"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12822567/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146032259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-01Epub Date: 2025-09-19DOI: 10.1007/978-3-032-04965-0_15
Minhui Yu, David S Lalush, Derek C Monroe, Kelly S Giovanello, Weili Lin, Pew-Thian Yap, Jason P Mihalik, Mingxia Liu
Multi-tracer positron emission tomography (PET), which assesses key neurological biomarkers such as tau pathology, neuroinflammatory, -amyloid deposition, and glucose metabolism, plays a vital role in diagnosing neurological disorders by providing complementary insights into the brain's molecular and functional state. Acquiring multi-tracer PET scans remains challenging due to high costs, radiation exposure, and limited tracer availability. Recent studies have attempted to synthesize multi-tracer PET images from structural MRI. However, these approaches typically either rely on direct mappings to individual tracers or lack distributional constraints, leading to inconsistencies in image quality across tracers. To this end, we propose a normalized diffusion framework (NDF) to generate high-quality multi-tracer PET images from a single MRI through a distribution-guided class-conditioned weighted diffusion model. Specifically, a diffusion model conditioned on MRI and tracer-specific class labels is trained to synthesize PET images of multiple tracers, and a pre-trained normalizing flow model refines these outputs by mapping them into a shared distribution space. This mapping ensures that the subject-specific high-level features across different PET tracers are preserved, resulting in more consistent and accurate synthesis. Experiments on a total of 425 subjects with multi-tracer PET scans demonstrate that our NDF outperforms current state-of-the-art methods, indicating its potential for advancing multi-tracer PET synthesis.
{"title":"Distribution-Guided Multi-Tracer Brain PET Synthesis from Structural MRI with Class-Conditioned Weighted Diffusion.","authors":"Minhui Yu, David S Lalush, Derek C Monroe, Kelly S Giovanello, Weili Lin, Pew-Thian Yap, Jason P Mihalik, Mingxia Liu","doi":"10.1007/978-3-032-04965-0_15","DOIUrl":"10.1007/978-3-032-04965-0_15","url":null,"abstract":"<p><p>Multi-tracer positron emission tomography (PET), which assesses key neurological biomarkers such as tau pathology, neuroinflammatory, <math><mi>β</mi></math> -amyloid deposition, and glucose metabolism, plays a vital role in diagnosing neurological disorders by providing complementary insights into the brain's molecular and functional state. Acquiring multi-tracer PET scans remains challenging due to high costs, radiation exposure, and limited tracer availability. Recent studies have attempted to synthesize multi-tracer PET images from structural MRI. However, these approaches typically either rely on direct mappings to individual tracers or lack distributional constraints, leading to inconsistencies in image quality across tracers. To this end, we propose a normalized diffusion framework (NDF) to generate high-quality multi-tracer PET images from a single MRI through a distribution-guided class-conditioned weighted diffusion model. Specifically, a diffusion model conditioned on MRI and tracer-specific class labels is trained to synthesize PET images of multiple tracers, and a pre-trained normalizing flow model refines these outputs by mapping them into a shared distribution space. This mapping ensures that the subject-specific high-level features across different PET tracers are preserved, resulting in more consistent and accurate synthesis. Experiments on a total of 425 subjects with multi-tracer PET scans demonstrate that our NDF outperforms current state-of-the-art methods, indicating its potential for advancing multi-tracer PET synthesis.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15963 ","pages":"152-162"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12747406/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145866927","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-01Epub Date: 2025-09-20DOI: 10.1007/978-3-032-04984-1_32
Jialu Pi, Juan Maria Farina, Rimita Lahiri, Jiwoong Jeong, Archana Gurudu, Hyung-Bok Park, Chieh-Ju Chao, Chadi Ayoub, Reza Arsanjani, Imon Banerjee
Major Adverse Cardiovascular Events (MACE) remain the leading cause of mortality globally, as reported in the Global Disease Burden Study 2021. Opportunistic screening leverages data collected from routine health check-ups and multimodal data can play a key role to identify at-risk individuals. Chest X-rays (CXR) provide insights into chronic conditions contributing to major adverse cardiovascular events (MACE), while 12-lead electrocardiogram (ECG) directly assesses cardiac electrical activity and structural abnormalities. Integrating CXR and ECG could offer a more comprehensive risk assessment than conventional models, which rely on clinical scores, computed tomography (CT) measurements, or biomarkers, which may be limited by sampling bias and single modality constraints. We propose a novel predictive modeling framework - MOSCARD, multimodal causal reasoning with co-attention to align two distinct modalities and simultaneously mitigate bias and confounders in opportunistic risk estimation. Primary technical contributions are - (i) multimodal alignment of CXR with ECG guidance; (ii) integration of causal reasoning; (iii) dual back-propagation graph for deconfounding. Evaluated on internal, shift data from emergency department (ED) and external MIMIC datasets, our model outperformed single (ED) and external MIMIC datasets, our model outperformed single modality and state-of-the-art foundational models - AUC: 0.75, 0.83, 0.71 respectively. Proposed cost-effective opportunistic screening enables early intervention, improving patient outcomes and reducing disparities.
{"title":"MOSCARD - Multimodal Opportunistic Screening for Cardiovascular Adverse events with Causal Reasoning and De-confounding.","authors":"Jialu Pi, Juan Maria Farina, Rimita Lahiri, Jiwoong Jeong, Archana Gurudu, Hyung-Bok Park, Chieh-Ju Chao, Chadi Ayoub, Reza Arsanjani, Imon Banerjee","doi":"10.1007/978-3-032-04984-1_32","DOIUrl":"10.1007/978-3-032-04984-1_32","url":null,"abstract":"<p><p>Major Adverse Cardiovascular Events (MACE) remain the leading cause of mortality globally, as reported in the Global Disease Burden Study 2021. Opportunistic screening leverages data collected from routine health check-ups and multimodal data can play a key role to identify at-risk individuals. Chest X-rays (CXR) provide insights into chronic conditions contributing to major adverse cardiovascular events (MACE), while 12-lead electrocardiogram (ECG) directly assesses cardiac electrical activity and structural abnormalities. Integrating CXR and ECG could offer a more comprehensive risk assessment than conventional models, which rely on clinical scores, computed tomography (CT) measurements, or biomarkers, which may be limited by sampling bias and single modality constraints. We propose a novel predictive modeling framework - MOSCARD, multimodal causal reasoning with co-attention to align two distinct modalities and simultaneously mitigate bias and confounders in opportunistic risk estimation. Primary technical contributions are - (i) multimodal alignment of CXR with ECG guidance; (ii) integration of causal reasoning; (iii) dual back-propagation graph for deconfounding. Evaluated on internal, shift data from emergency department (ED) and external MIMIC datasets, our model outperformed single (ED) and external MIMIC datasets, our model outperformed single modality and state-of-the-art foundational models - AUC: 0.75, 0.83, 0.71 respectively. Proposed cost-effective opportunistic screening enables early intervention, improving patient outcomes and reducing disparities.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15967 ","pages":"331-341"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12685155/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145716931","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-01Epub Date: 2025-09-18DOI: 10.1007/978-3-032-05182-0_37
Shuo Huang, Lujia Zhong, Yonggang Shi
For the early diagnosis of Alzheimer's disease (AD), it is essential that we have effective multiclass classification methods that can distinct subjects with mild cognitive impairment (MCI) from cognitively normal (CN) subjects and AD patients. However, significant overlaps of biomarker distributions among these groups make this a difficult task. In this work, we propose a novel framework for multi-modal, multiclass AD diagnosis that can integrate information from diverse and complex modalities to resolve ambiguity among the disease groups and hence enhance classification performances. More specifically, our approach integrates T1-weighted MRI, tau PET, fiber orientation distribution (FOD) from diffusion MRI (dMRI), and Montreal Cognitive Assessment (MoCA) scores to classify subjects into AD, MCI, and CN groups. We introduce a Swin-FOD model to extract order-balanced features from FOD and use contrastive learning to align MRI and PET features. These aligned features and MoCA scores are then processed with a Tabular Prior-data Fitted In-context Learning (TabPFN) method, which selects model parameters based on the alignment between input data and prior data during pre-training, eliminating the need for additional training or fine-tuning. Evaluated on the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset ( ), our model achieved a diagnosis accuracy of 73.21%, outperforming all comparison models ( ). We also performed Shapley analysis and quantitatively evaluated the essential contributions of each modality.
{"title":"Multistage Alignment and Fusion for Multimodal Multiclass Alzheimer's Disease Diagnosis.","authors":"Shuo Huang, Lujia Zhong, Yonggang Shi","doi":"10.1007/978-3-032-05182-0_37","DOIUrl":"10.1007/978-3-032-05182-0_37","url":null,"abstract":"<p><p>For the early diagnosis of Alzheimer's disease (AD), it is essential that we have effective multiclass classification methods that can distinct subjects with mild cognitive impairment (MCI) from cognitively normal (CN) subjects and AD patients. However, significant overlaps of biomarker distributions among these groups make this a difficult task. In this work, we propose a novel framework for multi-modal, multiclass AD diagnosis that can integrate information from diverse and complex modalities to resolve ambiguity among the disease groups and hence enhance classification performances. More specifically, our approach integrates T1-weighted MRI, tau PET, fiber orientation distribution (FOD) from diffusion MRI (dMRI), and Montreal Cognitive Assessment (MoCA) scores to classify subjects into AD, MCI, and CN groups. We introduce a Swin-FOD model to extract order-balanced features from FOD and use contrastive learning to align MRI and PET features. These aligned features and MoCA scores are then processed with a Tabular Prior-data Fitted In-context Learning (TabPFN) method, which selects model parameters based on the alignment between input data and prior data during pre-training, eliminating the need for additional training or fine-tuning. Evaluated on the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset ( <math><mi>n</mi> <mspace></mspace> <mo>=</mo> <mspace></mspace> <mn>1147</mn></math> ), our model achieved a diagnosis accuracy of 73.21%, outperforming all comparison models ( <math><mi>n</mi> <mo>=</mo> <mn>10</mn></math> ). We also performed Shapley analysis and quantitatively evaluated the essential contributions of each modality.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15974 ","pages":"375-385"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12674853/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145679841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-01Epub Date: 2025-09-21DOI: 10.1007/978-3-032-04947-6_2
Tong Chen, Minheng Chen, Jing Zhang, Yan Zhuang, Chao Cao, Xiaowei Yu, Yanjun Lyu, Lu Zhang, Li Su, Tianming Liu, Dajiang Zhu
Alzheimer's Disease (AD) and Lewy Body Dementia (LBD) often exhibit overlapping pathologies, leading to common symptoms that make diagnosis challenging and protracted in clinical settings. While many studies achieve promising accuracy in identifying AD and LBD at earlier stages, they often focus on discrete classification rather than capturing the gradual nature of disease progression. Since dementia develops progressively, understanding the continuous trajectory of dementia is crucial, as it allows us to uncover hidden patterns in cognitive decline and provides critical insights into the underlying mechanisms of disease progression. To address this gap, we propose a novel multi-scale learning framework that leverages hierarchical anatomical features to model the continuous relationships across various neurodegenerative conditions, including Mild Cognitive Impairment, AD, and LBD. Our approach employs the proposed hierarchical graph embedding fusion technique, integrating anatomical features, cortical folding patterns, and structural connectivity at multiple scales. This integration captures both fine-grained and coarse anatomical details, enabling the identification of subtle patterns that enhance differentiation between dementia types. Additionally, our framework projects each subject onto continuous tree structures, providing intuitive visualizations of disease trajectories and offering a more interpretable way to track cognitive decline. To validate our approach, we conduct extensive experiments on our in-house dataset of 308 subjects spanning multiple groups. Our results demonstrate that the proposed tree-based model effectively represents dementia progression, achieves promising performance in intricate classification task of AD and LBD, and highlights discriminative brain regions that contribute to the differentiation between dementia types. Our code is available at https://github.com/tongchen2010/haff.
{"title":"A Unified Continuous Staging Framework for Alzheimer's Disease and Lewy Body Dementia via Hierarchical Anatomical Features.","authors":"Tong Chen, Minheng Chen, Jing Zhang, Yan Zhuang, Chao Cao, Xiaowei Yu, Yanjun Lyu, Lu Zhang, Li Su, Tianming Liu, Dajiang Zhu","doi":"10.1007/978-3-032-04947-6_2","DOIUrl":"10.1007/978-3-032-04947-6_2","url":null,"abstract":"<p><p>Alzheimer's Disease (AD) and Lewy Body Dementia (LBD) often exhibit overlapping pathologies, leading to common symptoms that make diagnosis challenging and protracted in clinical settings. While many studies achieve promising accuracy in identifying AD and LBD at earlier stages, they often focus on discrete classification rather than capturing the gradual nature of disease progression. Since dementia develops progressively, understanding the continuous trajectory of dementia is crucial, as it allows us to uncover hidden patterns in cognitive decline and provides critical insights into the underlying mechanisms of disease progression. To address this gap, we propose a novel multi-scale learning framework that leverages hierarchical anatomical features to model the continuous relationships across various neurodegenerative conditions, including Mild Cognitive Impairment, AD, and LBD. Our approach employs the proposed hierarchical graph embedding fusion technique, integrating anatomical features, cortical folding patterns, and structural connectivity at multiple scales. This integration captures both fine-grained and coarse anatomical details, enabling the identification of subtle patterns that enhance differentiation between dementia types. Additionally, our framework projects each subject onto continuous tree structures, providing intuitive visualizations of disease trajectories and offering a more interpretable way to track cognitive decline. To validate our approach, we conduct extensive experiments on our in-house dataset of 308 subjects spanning multiple groups. Our results demonstrate that the proposed tree-based model effectively represents dementia progression, achieves promising performance in intricate classification task of AD and LBD, and highlights discriminative brain regions that contribute to the differentiation between dementia types. Our code is available at https://github.com/tongchen2010/haff.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15962 ","pages":"13-23"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12714422/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145807071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-01Epub Date: 2025-09-18DOI: 10.1007/978-3-032-05182-0_57
Jianwei Zhang, Yonggang Shi
Accurate modeling of disease progression is essential for comprehending the heterogeneous neuropathologies such as Alzheimer's Disease (AD). Traditional neuroimaging analysis often confound disease effects with normal aging, complicating the differential diagnosis. Recent advancements in deep learning have catalyzed the development of disentanglement techniques in Autoencoder networks, aiming to segregate longitudinal changes attributable to aging from those due to disease-specific alterations within the latent space. However, existing longitudinal disentanglement methods usually model disease as a single axis factor which ignores the complexity and heterogeneity of Alzheimer's Disease. In response to this issue, we propose a novel Surface-based Multi-axis Disentanglement framework.This framework posits multiple disease axes within the latent space, enhancing the model's capacity to encapsulate the multifaceted nature of AD, which includes various disease trajectories. To assign axes to data trajectories without explicit ground truth labels, we implement a longitudinal contrastive loss leveraging self-supervision, thereby refining the separation of disease trajectories. Evaluated on the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset ( ), our model demonstrates superior performance in delineating between cognitively normal (CN), mild cognitive impairment (MCI), and AD subjects,classification of stable MCI vs converting MCI and Amyloid status, compared to the single-axis model. This is further substantiated through an ablation study on the contrastive loss, underscoring the utility of our multi-axis approach in capturing the complex progression patterns of AD. The code is available at: https://github.com/jianweizhang17/MultiAxisDisentanglement.git.
{"title":"Surface-based Multi-Axis Longitudinal Disentanglement Using Contrastive Learning for Alzheimer's Disease.","authors":"Jianwei Zhang, Yonggang Shi","doi":"10.1007/978-3-032-05182-0_57","DOIUrl":"10.1007/978-3-032-05182-0_57","url":null,"abstract":"<p><p>Accurate modeling of disease progression is essential for comprehending the heterogeneous neuropathologies such as Alzheimer's Disease (AD). Traditional neuroimaging analysis often confound disease effects with normal aging, complicating the differential diagnosis. Recent advancements in deep learning have catalyzed the development of disentanglement techniques in Autoencoder networks, aiming to segregate longitudinal changes attributable to aging from those due to disease-specific alterations within the latent space. However, existing longitudinal disentanglement methods usually model disease as a single axis factor which ignores the complexity and heterogeneity of Alzheimer's Disease. In response to this issue, we propose a novel Surface-based Multi-axis Disentanglement framework.This framework posits multiple disease axes within the latent space, enhancing the model's capacity to encapsulate the multifaceted nature of AD, which includes various disease trajectories. To assign axes to data trajectories without explicit ground truth labels, we implement a longitudinal contrastive loss leveraging self-supervision, thereby refining the separation of disease trajectories. Evaluated on the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset ( <math><mi>N</mi> <mo>=</mo> <mn>1321</mn></math> ), our model demonstrates superior performance in delineating between cognitively normal (CN), mild cognitive impairment (MCI), and AD subjects,classification of stable MCI vs converting MCI and Amyloid status, compared to the single-axis model. This is further substantiated through an ablation study on the contrastive loss, underscoring the utility of our multi-axis approach in capturing the complex progression patterns of AD. The code is available at: https://github.com/jianweizhang17/MultiAxisDisentanglement.git.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15974 ","pages":"585-594"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12671042/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
During the process of brain aging, the changes of white matter structural connectivity are closely correlated with the cognitive traits and brain function. Genes have strong controls over this transition of structural connectivity-altering, which influences brain health and may lead to severe dementia disease, e.g., Alzheimer's disease. In this work, we introduce a novel deep-learning diagram, an oblique genomics mixture of experts(OG-MoE), designed to address the prediction of brain disease diagnosis, with awareness of the structural connectivity changes over time, and coupled with the genomics influences. By integrating genomics features into the dynamic gating router system of MoE layers, the model specializes in representing the structural connectivity components in separate parameter spaces. We pretrained the model on the self-regression task of brain connectivity predictions and then implemented multi-task supervised learning on brain disorder predictions and brain aging prediction. Compared to traditional associations analysis, this work provided a new way of discovering the soft but intricate inter-play between brain connectome phenotypes and genomic traits. It revealed the significant divergence of this correlation between the normal brain aging process and neurodegeneration.
{"title":"Oblique Genomics Mixture of Experts: Prediction of Brain Disorder With Aging-Related Changes of Brain's Structural Connectivity Under Genomic Influences.","authors":"Yanjun Lyu, Jing Zhang, Lu Zhang, Wei Ruan, Tianming Liu, Dajiang Zhu","doi":"10.1007/978-3-032-04965-0_33","DOIUrl":"10.1007/978-3-032-04965-0_33","url":null,"abstract":"<p><p>During the process of brain aging, the changes of white matter structural connectivity are closely correlated with the cognitive traits and brain function. Genes have strong controls over this transition of structural connectivity-altering, which influences brain health and may lead to severe dementia disease, e.g., Alzheimer's disease. In this work, we introduce a novel deep-learning diagram, an oblique genomics mixture of experts(OG-MoE), designed to address the prediction of brain disease diagnosis, with awareness of the structural connectivity changes over time, and coupled with the genomics influences. By integrating genomics features into the dynamic gating router system of MoE layers, the model specializes in representing the structural connectivity components in separate parameter spaces. We pretrained the model on the self-regression task of brain connectivity predictions and then implemented multi-task supervised learning on brain disorder predictions and brain aging prediction. Compared to traditional associations analysis, this work provided a new way of discovering the soft but intricate inter-play between brain connectome phenotypes and genomic traits. It revealed the significant divergence of this correlation between the normal brain aging process and neurodegeneration.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15963 ","pages":"348-358"},"PeriodicalIF":0.0,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12714490/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145807046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention