首页 > 最新文献

Computerized Medical Imaging and Graphics最新文献

英文 中文
CRAD: Cognitive Aware Feature Refinement with Missing Modalities for Early Alzheimer’s Progression Prediction 认知意识特征细化与缺失模式在早期阿尔茨海默病进展预测中的应用。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-11-10 DOI: 10.1016/j.compmedimag.2025.102664
Fei Liu , Shiuan-Ni Liang , Mohamed Hisham Jaward , Huey Fang Ong , Huabin Wang , Alzheimer’s Disease Neuroimaging Initiative , Australian Imaging Biomarkers and Lifestyle flagship study of ageing
Accurate diagnosis and early prediction of Alzheimer’s disease (AD) often require multiple neuroimageing modalities, but in many cases, only one or two modalities are available. This missing modality hinders the accuracy of diagnosis and is a critical challenge in clinical practice. Multimodal knowledge distillation (KD) offers a promising solution by aligning complete knowledge from multimodal data with that of partial modalities. However, current methods focus on aligning high-level features, which limit their effectiveness due to insufficient transfer of reliable knowledge. In this work, we propose a novel Consistency Refinement-driven Multi-level Self-Attention Distillation framework (CRAD) for Early Alzheimer’s Progression Prediction, which enables the cross-modal transfer of more robust shallow knowledge with self-attention to refine features. We develop a multi-level distillation module to progressively distill cross-modal discriminating knowledge, enabling lightweight yet reliable knowledge transfer. Moreover, we design a novel self-attention distillation module (PF-CMAD) to transfer disease-relevant intermediate knowledge, which leverages feature self-similarity to capture cross-modal correlations without introducing trainable parameters, enabling interpretable and efficient distillation. We incorporate a consistency-evaluation-driven confidence regularization strategy within the distillation process. This strategy dynamically refines knowledge using adaptive distillation controllers that assess teacher confidence. Comprehensive experiments demonstrate that our method achieves superior accuracy and robust cross-dataset generalization performance using only MRI for AD diagnosis and early progression prediction. The code is available at https://github.com/LiuFei-AHU/CRAD.
阿尔茨海默病(AD)的准确诊断和早期预测通常需要多种神经成像方式,但在许多情况下,只有一种或两种方式可用。这种缺失的模式阻碍了诊断的准确性,是临床实践中的一个关键挑战。多模态知识蒸馏(KD)通过将来自多模态数据的完整知识与部分模态的知识进行比对,提供了一种很有前途的解决方案。然而,目前的方法主要集中在高级特征的对齐上,由于可靠知识的转移不足,限制了它们的有效性。在这项工作中,我们提出了一种新的一致性改进驱动的多层次自关注蒸馏框架(CRAD)用于早期阿尔茨海默氏症的进展预测,该框架能够跨模态转移更强大的具有自关注的浅层知识来改进特征。我们开发了一个多级蒸馏模块,逐步提取跨模态的鉴别知识,实现轻量级但可靠的知识转移。此外,我们设计了一种新的自关注蒸馏模块(PF-CMAD)来转移疾病相关的中间知识,该模块利用特征自相似性来捕获跨模态相关性,而不引入可训练的参数,从而实现可解释和高效的蒸馏。我们在蒸馏过程中加入了一致性评估驱动的置信度正则化策略。该策略使用评估教师信心的自适应蒸馏控制器动态地提炼知识。综合实验表明,该方法仅使用MRI进行AD诊断和早期进展预测,具有优越的准确性和鲁棒的跨数据集泛化性能。代码可在https://github.com/LiuFei-AHU/CRAD上获得。
{"title":"CRAD: Cognitive Aware Feature Refinement with Missing Modalities for Early Alzheimer’s Progression Prediction","authors":"Fei Liu ,&nbsp;Shiuan-Ni Liang ,&nbsp;Mohamed Hisham Jaward ,&nbsp;Huey Fang Ong ,&nbsp;Huabin Wang ,&nbsp;Alzheimer’s Disease Neuroimaging Initiative ,&nbsp;Australian Imaging Biomarkers and Lifestyle flagship study of ageing","doi":"10.1016/j.compmedimag.2025.102664","DOIUrl":"10.1016/j.compmedimag.2025.102664","url":null,"abstract":"<div><div>Accurate diagnosis and early prediction of Alzheimer’s disease (AD) often require multiple neuroimageing modalities, but in many cases, only one or two modalities are available. This missing modality hinders the accuracy of diagnosis and is a critical challenge in clinical practice. Multimodal knowledge distillation (KD) offers a promising solution by aligning complete knowledge from multimodal data with that of partial modalities. However, current methods focus on aligning high-level features, which limit their effectiveness due to insufficient transfer of reliable knowledge. In this work, we propose a novel Consistency Refinement-driven Multi-level Self-Attention Distillation framework (CRAD) for Early Alzheimer’s Progression Prediction, which enables the cross-modal transfer of more robust shallow knowledge with self-attention to refine features. We develop a multi-level distillation module to progressively distill cross-modal discriminating knowledge, enabling lightweight yet reliable knowledge transfer. Moreover, we design a novel self-attention distillation module (PF-CMAD) to transfer disease-relevant intermediate knowledge, which leverages feature self-similarity to capture cross-modal correlations without introducing trainable parameters, enabling interpretable and efficient distillation. We incorporate a consistency-evaluation-driven confidence regularization strategy within the distillation process. This strategy dynamically refines knowledge using adaptive distillation controllers that assess teacher confidence. Comprehensive experiments demonstrate that our method achieves superior accuracy and robust cross-dataset generalization performance using only MRI for AD diagnosis and early progression prediction. The code is available at <span><span>https://github.com/LiuFei-AHU/CRAD</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102664"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145507841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A CNN-Transformer fusion network for Diabetic retinopathy image classification 用于糖尿病视网膜病变图像分类的CNN-Transformer融合网络。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-10-21 DOI: 10.1016/j.compmedimag.2025.102655
Xuan Huang , Zhuang Ai , Chongyang She , Qi Li , Qihao Wei , Sha Xu , Yaping Lu , Fanxin Zeng
Diabetic retinopathy (DR) is a leading cause of blindness worldwide, yet current diagnosis relies on labor-intensive and subjective fundus image interpretation. Here we present a convolutional neural network-transformer fusion model (DR-CTFN) that integrates ConvNeXt and Swin Transformer algorithms with a lightweight attention block (LAB) to enhance feature extraction. To address dataset imbalance, we applied standardized preprocessing and extensive image augmentation. On the Kaggle EyePACS dataset, DR-CTFN outperformed ConvNeXt and Swin Transformer in accuracy by 3.14% and 8.39%, while also achieving a superior area under the curve (AUC) by 1% and 26.08%. External validation on APTOS 2019 Blindness Detection and a clinical DR dataset yielded accuracies of 84.45% and 85.31%, with AUC values of 95.22% and 95.79%, respectively. These results demonstrate that DR-CTFN enables rapid, robust, and precise DR detection, offering a scalable approach for early diagnosis and prevention of vision loss, thereby enhancing the quality of life for DR patients.
糖尿病视网膜病变(DR)是世界范围内致盲的主要原因,但目前的诊断依赖于劳动密集型和主观眼底图像解释。本文提出了一种卷积神经网络-变压器融合模型(DR-CTFN),该模型将ConvNeXt和Swin Transformer算法与轻量级注意块(LAB)集成在一起,以增强特征提取。为了解决数据不平衡问题,我们采用了标准化的预处理和广泛的图像增强。在Kaggle EyePACS数据集上,DR-CTFN的准确率分别比ConvNeXt和Swin Transformer高3.14%和8.39%,同时曲线下面积(AUC)也比ConvNeXt和Swin Transformer高1%和26.08%。在APTOS 2019盲目性检测和临床DR数据集上进行外部验证,准确率分别为84.45%和85.31%,AUC值分别为95.22%和95.79%。这些结果表明,DR- ctfn能够实现快速、稳健和精确的DR检测,为早期诊断和预防视力丧失提供了一种可扩展的方法,从而提高了DR患者的生活质量。
{"title":"A CNN-Transformer fusion network for Diabetic retinopathy image classification","authors":"Xuan Huang ,&nbsp;Zhuang Ai ,&nbsp;Chongyang She ,&nbsp;Qi Li ,&nbsp;Qihao Wei ,&nbsp;Sha Xu ,&nbsp;Yaping Lu ,&nbsp;Fanxin Zeng","doi":"10.1016/j.compmedimag.2025.102655","DOIUrl":"10.1016/j.compmedimag.2025.102655","url":null,"abstract":"<div><div>Diabetic retinopathy (DR) is a leading cause of blindness worldwide, yet current diagnosis relies on labor-intensive and subjective fundus image interpretation. Here we present a convolutional neural network-transformer fusion model (DR-CTFN) that integrates ConvNeXt and Swin Transformer algorithms with a lightweight attention block (LAB) to enhance feature extraction. To address dataset imbalance, we applied standardized preprocessing and extensive image augmentation. On the Kaggle EyePACS dataset, DR-CTFN outperformed ConvNeXt and Swin Transformer in accuracy by 3.14% and 8.39%, while also achieving a superior area under the curve (AUC) by 1% and 26.08%. External validation on APTOS 2019 Blindness Detection and a clinical DR dataset yielded accuracies of 84.45% and 85.31%, with AUC values of 95.22% and 95.79%, respectively. These results demonstrate that DR-CTFN enables rapid, robust, and precise DR detection, offering a scalable approach for early diagnosis and prevention of vision loss, thereby enhancing the quality of life for DR patients.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102655"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145394925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ESAM2-BLS: Enhanced segment anything model 2 for efficient breast lesion segmentation in ultrasound imaging ESAM2-BLS:用于超声成像中乳腺病变有效分割的增强分段任何模型2。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-10-10 DOI: 10.1016/j.compmedimag.2025.102654
Lishuang Guo , Haonan Zhang , Chenbin Ma
Ultrasound imaging, as an economical, efficient, and non-invasive diagnostic tool, is widely used for breast lesion screening and diagnosis. However, the segmentation of lesion regions remains a significant challenge due to factors such as noise interference and the variability in image quality. To address this issue, we propose a novel deep learning model named enhanced segment anything model 2 (SAM2) for breast lesion segmentation (ESAM2-BLS). This model is an optimized version of the SAM2 architecture. ESAM2-BLS customizes and fine-tunes the pre-trained SAM2 model by introducing an adapter module, specifically designed to accommodate the unique characteristics of breast ultrasound images. The adapter module directly addresses ultrasound-specific challenges including speckle noise, low contrast boundaries, shadowing artifacts, and anisotropic resolution through targeted architectural elements such as channel attention mechanisms, specialized convolution kernels, and optimized skip connections. This optimization significantly improves segmentation accuracy, particularly for low-contrast and small lesion regions. Compared to traditional methods, ESAM2-BLS fully leverages the generalization capabilities of large models while incorporating multi-scale feature fusion and axial dilated depthwise convolution to effectively capture multi-level information from complex lesions. During the decoding process, the model enhances the identification of fine boundaries and small lesions through depthwise separable convolutions and skip connections, while maintaining a low computational cost. Visualization of the segmentation results and interpretability analysis demonstrate that ESAM2-BLS achieves an average Dice score of 0.9077 and 0.8633 in five-fold cross-validation across two datasets with over 1600 patients. These results significantly improve segmentation accuracy and robustness. This model provides an efficient, reliable, and specialized automated solution for early breast cancer screening and diagnosis.
超声成像作为一种经济、高效、无创的诊断手段,被广泛应用于乳腺病变的筛查和诊断。然而,由于噪声干扰和图像质量的可变性等因素,病灶区域的分割仍然是一个重大挑战。为了解决这个问题,我们提出了一种新的深度学习模型,称为增强分割任何模型2 (SAM2),用于乳腺病变分割(ESAM2-BLS)。该模型是SAM2体系结构的优化版本。ESAM2-BLS定制和微调预训练SAM2模型通过引入一个适配器模块,专门设计以适应乳房超声图像的独特特点。适配器模块通过通道注意机制、专门的卷积核和优化的跳过连接等目标架构元素,直接解决超声波特定的挑战,包括散斑噪声、低对比度边界、阴影伪影和各向异性分辨率。这种优化显著提高了分割精度,特别是对于低对比度和小病变区域。与传统方法相比,ESAM2-BLS充分利用了大型模型的泛化能力,同时结合了多尺度特征融合和轴向扩张深度卷积,有效地捕获了复杂病变的多层次信息。在解码过程中,该模型通过深度可分离卷积和跳跃连接增强了对细边界和小病灶的识别,同时保持了较低的计算成本。分割结果的可视化和可解释性分析表明,ESAM2-BLS在超过1600例患者的两个数据集上进行了五倍交叉验证,平均Dice得分为0.9077和0.8633。这些结果显著提高了分割的准确性和鲁棒性。该模型为早期乳腺癌筛查和诊断提供了高效、可靠、专业化的自动化解决方案。
{"title":"ESAM2-BLS: Enhanced segment anything model 2 for efficient breast lesion segmentation in ultrasound imaging","authors":"Lishuang Guo ,&nbsp;Haonan Zhang ,&nbsp;Chenbin Ma","doi":"10.1016/j.compmedimag.2025.102654","DOIUrl":"10.1016/j.compmedimag.2025.102654","url":null,"abstract":"<div><div>Ultrasound imaging, as an economical, efficient, and non-invasive diagnostic tool, is widely used for breast lesion screening and diagnosis. However, the segmentation of lesion regions remains a significant challenge due to factors such as noise interference and the variability in image quality. To address this issue, we propose a novel deep learning model named enhanced segment anything model 2 (SAM2) for breast lesion segmentation (ESAM2-BLS). This model is an optimized version of the SAM2 architecture. ESAM2-BLS customizes and fine-tunes the pre-trained SAM2 model by introducing an adapter module, specifically designed to accommodate the unique characteristics of breast ultrasound images. The adapter module directly addresses ultrasound-specific challenges including speckle noise, low contrast boundaries, shadowing artifacts, and anisotropic resolution through targeted architectural elements such as channel attention mechanisms, specialized convolution kernels, and optimized skip connections. This optimization significantly improves segmentation accuracy, particularly for low-contrast and small lesion regions. Compared to traditional methods, ESAM2-BLS fully leverages the generalization capabilities of large models while incorporating multi-scale feature fusion and axial dilated depthwise convolution to effectively capture multi-level information from complex lesions. During the decoding process, the model enhances the identification of fine boundaries and small lesions through depthwise separable convolutions and skip connections, while maintaining a low computational cost. Visualization of the segmentation results and interpretability analysis demonstrate that ESAM2-BLS achieves an average Dice score of 0.9077 and 0.8633 in five-fold cross-validation across two datasets with over 1600 patients. These results significantly improve segmentation accuracy and robustness. This model provides an efficient, reliable, and specialized automated solution for early breast cancer screening and diagnosis.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102654"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145356747","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multistain multicompartment automatic segmentation in renal biopsies with thrombotic microangiopathies and other vasculopathies 血栓性微血管病变和其他血管病变肾活检的多染色多室自动分割。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-10-22 DOI: 10.1016/j.compmedimag.2025.102658
Nicola Altini , Michela Prunella , Surya V. Seshan , Savino Sciascia , Antonella Barreca , Alessandro Del Gobbo , Stefan Porubsky , Hien Van Nguyen , Claudia Delprete , Berardino Prencipe , Deján Dobi , Daan P.C. van Doorn , Sjoerd A.M.E.G. Timmermans , Pieter van Paassen , Vitoantonio Bevilacqua , Jan Ulrich Becker
Automatic tissue segmentation is a necessary step for the bulk analysis of whole slide images (WSIs) from paraffin histology sections in kidney biopsies. However, existing models often fail to generalize across the main nephropathological staining methods and to capture the severe morphological distortions in arteries, arterioles, and glomeruli common in thrombotic microangiopathy (TMA) or other vasculopathies. Therefore, we developed an automatic multi-staining segmentation pipeline covering six key compartments: Artery, Arteriole, Glomerulus, Cortex, Medulla, and Capsule/Other. This framework enables downstream tasks such as counting and labeling at instance-, WSI- or biopsy-level. Biopsies (n = 158) from seven centers: Cologne, Turin, Milan, Weill-Cornell, Mainz, Maastricht, Budapest, were classified by expert nephropathologists into TMA (n = 87) or Mimickers (n = 71). Ground truth expert segmentation masks were provided for all compartments, and expert binary TMA classification labels for Glomerulus, Artery, Arteriole. The biopsies were divided into training (n = 79), validation (n = 26), and test (n = 53) subsets. We benchmarked six deep learning models for semantic segmentation (U-Net, FPN, DeepLabV3+, Mask2Former, SegFormer, SegNeXt) and five models for classification (ResNet-34, DenseNet-121, EfficientNet-v2-S, ConvNeXt-Small, Swin-v2-B). We obtained robust segmentation results across all compartments. On the test set, the best models achieved Dice coefficients of 0.903 (Cortex), 0.834 (Medulla), 0.816 (Capsule/Other), 0.922 (Glomerulus), 0.822 (Artery), and 0.553 (Arteriole). The best classification models achieved Accuracy of 0.724 and 0.841 for Glomerulus and Artery plus Arteriole compartments, respectively. Furthermore, we release NePathTK (NephroPathology Toolkit), a powerful open-source end-to-end pipeline integrated with QuPath, enabling accurate segmentation for decision support in nephropathology and large-scale analysis of kidney biopsies.
自动组织分割是整个切片图像(wsi)的散装分析的必要步骤,从石蜡组织切片肾活检。然而,现有的模型往往不能概括主要的肾脏病理染色方法,也不能捕捉血栓性微血管病(TMA)或其他血管病变中常见的动脉、小动脉和肾小球的严重形态扭曲。因此,我们开发了一种自动多染色分割管道,涵盖六个关键区室:动脉、小动脉、肾小球、皮质、髓质和胶囊/其他。该框架支持下游任务,例如实例级、WSI级或活检级的计数和标记。来自科隆、都灵、米兰、威尔-康奈尔、美因茨、马斯特里赫特、布达佩斯七个中心的活检(n = 158)由肾病理学专家分为TMA (n = 87)或Mimickers (n = 71)。为所有隔室提供了Ground truth专家分割掩码,并为肾小球、动脉、小动脉提供了专家二元TMA分类标签。将活检分为训练组(n = 79)、验证组(n = 26)和测试组(n = 53)。我们对语义分割的6个深度学习模型(U-Net、FPN、DeepLabV3+、Mask2Former、SegFormer、SegNeXt)和分类的5个模型(ResNet-34、DenseNet-121、EfficientNet-v2-S、ConvNeXt-Small、swun -v2- b)进行了基准测试。我们在所有隔室中获得了稳健的分割结果。在测试集上,最佳模型的Dice系数分别为0.903(皮质)、0.834(髓质)、0.816(胶囊/其他)、0.922(肾小球)、0.822(动脉)和0.553(动脉)。最佳分类模型对肾小球室和动脉+小动脉室的准确率分别为0.724和0.841。此外,我们还发布了NePathTK(肾脏病理学工具包),这是一个强大的开源端到端管道,与QuPath集成,可以为肾脏病理学决策支持和肾脏活检的大规模分析提供准确的分割。
{"title":"Multistain multicompartment automatic segmentation in renal biopsies with thrombotic microangiopathies and other vasculopathies","authors":"Nicola Altini ,&nbsp;Michela Prunella ,&nbsp;Surya V. Seshan ,&nbsp;Savino Sciascia ,&nbsp;Antonella Barreca ,&nbsp;Alessandro Del Gobbo ,&nbsp;Stefan Porubsky ,&nbsp;Hien Van Nguyen ,&nbsp;Claudia Delprete ,&nbsp;Berardino Prencipe ,&nbsp;Deján Dobi ,&nbsp;Daan P.C. van Doorn ,&nbsp;Sjoerd A.M.E.G. Timmermans ,&nbsp;Pieter van Paassen ,&nbsp;Vitoantonio Bevilacqua ,&nbsp;Jan Ulrich Becker","doi":"10.1016/j.compmedimag.2025.102658","DOIUrl":"10.1016/j.compmedimag.2025.102658","url":null,"abstract":"<div><div>Automatic tissue segmentation is a necessary step for the bulk analysis of whole slide images (WSIs) from paraffin histology sections in kidney biopsies. However, existing models often fail to generalize across the main nephropathological staining methods and to capture the severe morphological distortions in arteries, arterioles, and glomeruli common in thrombotic microangiopathy (TMA) or other vasculopathies. Therefore, we developed an automatic multi-staining segmentation pipeline covering six key compartments: Artery, Arteriole, Glomerulus, Cortex, Medulla, and Capsule/Other. This framework enables downstream tasks such as counting and labeling at instance-, WSI- or biopsy-level. Biopsies (n = 158) from seven centers: Cologne, Turin, Milan, Weill-Cornell, Mainz, Maastricht, Budapest, were classified by expert nephropathologists into TMA (n = 87) or Mimickers (n = 71). Ground truth expert segmentation masks were provided for all compartments, and expert binary TMA classification labels for Glomerulus, Artery, Arteriole. The biopsies were divided into training (n = 79), validation (n = 26), and test (n = 53) subsets. We benchmarked six deep learning models for semantic segmentation (U-Net, FPN, DeepLabV3+, Mask2Former, SegFormer, SegNeXt) and five models for classification (ResNet-34, DenseNet-121, EfficientNet-v2-S, ConvNeXt-Small, Swin-v2-B). We obtained robust segmentation results across all compartments. On the test set, the best models achieved Dice coefficients of 0.903 (Cortex), 0.834 (Medulla), 0.816 (Capsule/Other), 0.922 (Glomerulus), 0.822 (Artery), and 0.553 (Arteriole). The best classification models achieved Accuracy of 0.724 and 0.841 for Glomerulus and Artery plus Arteriole compartments, respectively. Furthermore, we release NePathTK (NephroPathology Toolkit), a powerful open-source end-to-end pipeline integrated with QuPath, enabling accurate segmentation for decision support in nephropathology and large-scale analysis of kidney biopsies.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102658"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145410649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient frequency-decomposed transformer via large vision model guidance for surgical image desmoking 基于大视觉模型引导的高效分频变压器用于手术图像去噪
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-11-03 DOI: 10.1016/j.compmedimag.2025.102660
Jiaao Li , Diandian Guo , Youyu Wang , Yanhui Wan , Long Ma , Jialun Pei
Surgical image restoration plays a vital clinical role in improving visual quality during surgery, particularly in minimally invasive procedures where the operating field is frequently obscured by surgical smoke. However, surgical image desmoking still has limited progress in algorithm development and customized learning strategies. In this regard, this work focuses on the task of desmoking from both theoretical and practical perspectives. First, we analyze the intrinsic characteristics of surgical smoke degradation: (1) spatial localization and dynamics, (2) distinguishable frequency-domain patterns, and (3) the entangled representation of anatomical content and degradative artifacts. These observations motivated us to propose an efficient frequency-aware Transformer framework, namely SmoRestor, which aims to separate and restore true anatomical structures from complex degradations. Specifically, we introduce a high-order Fourier-embedded neighborhood attention transformer that enhances the model’s ability to capture structured degradation patterns across both spatial and frequency domains. Besides, we utilize the semantic priors encoded by large vision models to disambiguate content from degradation through targeted guidance. Moreover, we propose an innovative transfer learning paradigm that injects knowledge from large models to the main network, enabling it to effectively distinguish meaningful content from ambiguous corruption. Experimental results on both public and in-house datasets demonstrate substantial improvements in quantitative performance and visual quality. The source code will be available.
手术图像恢复在提高手术视觉质量方面起着至关重要的临床作用,特别是在微创手术中,手术视野经常被手术烟雾遮挡。然而,手术图像去吸烟在算法开发和定制学习策略方面仍然进展有限。在这方面,本工作着重从理论和实践两个角度来研究吸烟的任务。首先,我们分析了手术烟雾降解的内在特征:(1)空间定位和动力学;(2)可区分的频域模式;(3)解剖内容和降解伪像的纠缠表示。这些观察促使我们提出了一种有效的频率感知变压器框架,即SmoRestor,旨在从复杂的退化中分离和恢复真实的解剖结构。具体来说,我们引入了一个高阶傅里叶嵌入式邻域注意力转换器,增强了模型在空间和频域捕获结构化退化模式的能力。此外,我们利用大视觉模型编码的语义先验,通过有针对性的引导来消除内容的歧义。此外,我们提出了一种创新的迁移学习范式,将大型模型中的知识注入主网络,使其能够有效区分有意义的内容和模糊的腐败内容。在公共和内部数据集上的实验结果表明,在定量性能和视觉质量方面有了实质性的改进。源代码将可用。
{"title":"Efficient frequency-decomposed transformer via large vision model guidance for surgical image desmoking","authors":"Jiaao Li ,&nbsp;Diandian Guo ,&nbsp;Youyu Wang ,&nbsp;Yanhui Wan ,&nbsp;Long Ma ,&nbsp;Jialun Pei","doi":"10.1016/j.compmedimag.2025.102660","DOIUrl":"10.1016/j.compmedimag.2025.102660","url":null,"abstract":"<div><div>Surgical image restoration plays a vital clinical role in improving visual quality during surgery, particularly in minimally invasive procedures where the operating field is frequently obscured by surgical smoke. However, surgical image desmoking still has limited progress in algorithm development and customized learning strategies. In this regard, this work focuses on the task of desmoking from both theoretical and practical perspectives. First, we analyze the intrinsic characteristics of surgical smoke degradation: (1) spatial localization and dynamics, (2) distinguishable frequency-domain patterns, and (3) the entangled representation of anatomical content and degradative artifacts. These observations motivated us to propose an efficient frequency-aware Transformer framework, namely SmoRestor, which aims to separate and restore true anatomical structures from complex degradations. Specifically, we introduce a high-order Fourier-embedded neighborhood attention transformer that enhances the model’s ability to capture structured degradation patterns across both spatial and frequency domains. Besides, we utilize the semantic priors encoded by large vision models to disambiguate content from degradation through targeted guidance. Moreover, we propose an innovative transfer learning paradigm that injects knowledge from large models to the main network, enabling it to effectively distinguish meaningful content from ambiguous corruption. Experimental results on both public and in-house datasets demonstrate substantial improvements in quantitative performance and visual quality. The source code will be available.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102660"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145466996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Path and bone-contour regularized unpaired MRI-to-CT translation 路径和骨轮廓正则化非配对mri - ct翻译
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-10-13 DOI: 10.1016/j.compmedimag.2025.102656
Teng Zhou , Jax Luo , Yuping Sun , Yiheng Tan , Shun Yao , Nazim Haouchine , Scott Raymond
Accurate MRI-to-CT translation promises the integration of complementary imaging information without the need for additional imaging sessions. Given the practical challenges associated with acquiring paired MRI and CT scans, the development of robust methods capable of leveraging unpaired datasets is essential for advancing the MRI-to-CT translation. Current unpaired MRI-to-CT translation methods, which predominantly rely on cycle consistency and contrastive learning frameworks, frequently encounter challenges in accurately translating anatomical features that are highly discernible on CT but less distinguishable on MRI, such as bone structures. This limitation renders these approaches less suitable for applications in radiation therapy, where precise bone representation is essential for accurate treatment planning. To address this challenge, we propose a path- and bone-contour regularized approach for unpaired MRI-to-CT translation. In our method, MRI and CT images are projected to a shared latent space, where the MRI-to-CT mapping is modeled as a continuous flow governed by neural ordinary differential equations. The optimal mapping is obtained by minimizing the transition path length of the flow. To enhance the accuracy of translated bone structures, we introduce a trainable neural network to generate bone contours from MRI and implement mechanisms to directly and indirectly encourage the model to focus on bone contours and their adjacent regions. Evaluations conducted on three datasets demonstrate that our method outperforms existing unpaired MRI-to-CT translation approaches, achieving lower overall error rates. Moreover, in a downstream bone segmentation task, our approach exhibits superior performance in preserving the fidelity of bone structures. Our code is available at: https://github.com/kennysyp/PaBoT.
准确的mri到ct转换保证了互补成像信息的整合,而不需要额外的成像会话。考虑到获取配对MRI和CT扫描相关的实际挑战,开发能够利用非配对数据集的强大方法对于推进MRI到CT的转换至关重要。目前的非配对MRI- CT翻译方法主要依赖于周期一致性和对比学习框架,在准确翻译在CT上高度可识别但在MRI上不易识别的解剖特征(如骨结构)时经常遇到挑战。这种限制使得这些方法不太适合应用于放射治疗,在放射治疗中,精确的骨表示对于准确的治疗计划至关重要。为了解决这一挑战,我们提出了一种非配对mri到ct翻译的路径和骨轮廓正则化方法。在我们的方法中,MRI和CT图像被投影到一个共享的潜在空间,其中MRI到CT的映射被建模为由神经常微分方程控制的连续流。通过最小化流的过渡路径长度来获得最优映射。为了提高翻译骨结构的准确性,我们引入了一个可训练的神经网络来从MRI中生成骨轮廓,并实现了直接和间接鼓励模型关注骨轮廓及其邻近区域的机制。在三个数据集上进行的评估表明,我们的方法优于现有的非成对mri - ct翻译方法,实现了更低的总体错误率。此外,在下游的骨分割任务中,我们的方法在保持骨结构的保真度方面表现出优越的性能。我们的代码可在:https://github.com/kennysyp/PaBoT。
{"title":"Path and bone-contour regularized unpaired MRI-to-CT translation","authors":"Teng Zhou ,&nbsp;Jax Luo ,&nbsp;Yuping Sun ,&nbsp;Yiheng Tan ,&nbsp;Shun Yao ,&nbsp;Nazim Haouchine ,&nbsp;Scott Raymond","doi":"10.1016/j.compmedimag.2025.102656","DOIUrl":"10.1016/j.compmedimag.2025.102656","url":null,"abstract":"<div><div>Accurate MRI-to-CT translation promises the integration of complementary imaging information without the need for additional imaging sessions. Given the practical challenges associated with acquiring paired MRI and CT scans, the development of robust methods capable of leveraging unpaired datasets is essential for advancing the MRI-to-CT translation. Current unpaired MRI-to-CT translation methods, which predominantly rely on cycle consistency and contrastive learning frameworks, frequently encounter challenges in accurately translating anatomical features that are highly discernible on CT but less distinguishable on MRI, such as bone structures. This limitation renders these approaches less suitable for applications in radiation therapy, where precise bone representation is essential for accurate treatment planning. To address this challenge, we propose a path- and bone-contour regularized approach for unpaired MRI-to-CT translation. In our method, MRI and CT images are projected to a shared latent space, where the MRI-to-CT mapping is modeled as a continuous flow governed by neural ordinary differential equations. The optimal mapping is obtained by minimizing the transition path length of the flow. To enhance the accuracy of translated bone structures, we introduce a trainable neural network to generate bone contours from MRI and implement mechanisms to directly and indirectly encourage the model to focus on bone contours and their adjacent regions. Evaluations conducted on three datasets demonstrate that our method outperforms existing unpaired MRI-to-CT translation approaches, achieving lower overall error rates. Moreover, in a downstream bone segmentation task, our approach exhibits superior performance in preserving the fidelity of bone structures. Our code is available at: <span><span>https://github.com/kennysyp/PaBoT</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102656"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145290008","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Colorectal disease diagnosis with deep triple-stream fusion and attention refinement 结直肠疾病的深三流融合与注意精细化诊断
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-11-15 DOI: 10.1016/j.compmedimag.2025.102669
Abdulfattah Ba Alawi , Abdullah Ammar Karcioglu , Ferhat Bozkurt
Colorectal cancer constitutes a significant proportion of global cancer-related mortality, underscoring the imperative for robust and early-stage diagnostic methodologies. In this study, we propose a novel end-to-end deep learning framework that integrates multiple advanced mechanisms to enhance the classification of colorectal disease from histopathologic and endoscopic images. Our model, named TripleFusionNet, leverages a unique triple-stream architecture by combining the strengths of EfficientNetB3, ResNet50, and DenseNet121, enabling the extraction of rich, multi-level feature representations from input images. To augment discriminative feature modeling, a Multi-Scale Attention Module is integrated, which concurrently performs spatial and channel-wise recalibration, thereby enabling the network to emphasize diagnostically salient regions. Additionally, we incorporate a Squeeze-Excite Refinement Block (SERB) to selectively enhance informative channel activations while attenuating noise and redundant signals. Feature representations from the individual backbones are adaptively fused through a Progressive Gated Fusion mechanism that dynamically learns context-aware weighting for optimal feature integration and redundancy mitigation. We validate our approach on two colorectal benchmarks: CRCCD_V1 (14 classes) and LC25000 (binary). On CRCCD_V1, the best performance is obtained by a conventional classifier trained on our 256-D TripleFusionNet embeddings—SVM (RBF) reaches 96.63% test accuracy with macro F1 96.62%, with the Stacking Ensemble close behind. With five-fold cross-validation, it yields comparable out-of-fold means (0.964 with small standard deviations), confirming stability across partitions. End-to-end image-based baselines, including TripleFusionNet, are competitive but are slightly surpassed by embedding-based classifiers, highlighting the utility of the learned representation. On LC25000, our method attains 100% accuracy. Beyond accuracy, the approach maintains strong precision, recall, F1, and ROC–AUC, and the fused embeddings transfer effectively to multiple conventional learners (e.g., Random Forest, XGBoost). These results confirm the potential of the model for real-world deployment in computer-aided diagnosis workflows, particularly within resource-constrained clinical settings.
结直肠癌在全球癌症相关死亡率中占很大比例,因此需要强有力的早期诊断方法。在这项研究中,我们提出了一个新的端到端深度学习框架,该框架集成了多种先进的机制,以增强从组织病理学和内窥镜图像中对结直肠疾病的分类。我们的模型名为TripleFusionNet,通过结合EfficientNetB3、ResNet50和DenseNet121的优势,利用独特的三流架构,能够从输入图像中提取丰富的、多层次的特征表示。为了增强判别特征建模,集成了一个多尺度注意力模块,该模块同时执行空间和通道重新校准,从而使网络能够强调诊断上的显著区域。此外,我们还结合了一个压缩激励细化块(塞族),以选择性地增强信息通道激活,同时衰减噪声和冗余信号。来自各个主干网的特征表示通过渐进门控融合机制自适应融合,该机制动态学习上下文感知权重,以实现最佳特征集成和冗余缓解。我们在两个结肠直肠基准上验证了我们的方法:CRCCD_V1(14类)和LC25000(二进制)。在CRCCD_V1上,在我们的256-D TripleFusionNet嵌入上训练的传统分类器-支持向量机(RBF)的测试准确率达到96.63%,宏F1为96.62%,Stacking Ensemble紧随其后。通过五倍交叉验证,它产生了可比较的叠外均值(0.964,标准差较小),确认了跨分区的稳定性。端到端基于图像的基线,包括TripleFusionNet,是有竞争力的,但被基于嵌入的分类器稍微超越,突出了学习表征的实用性。在LC25000上,我们的方法达到了100%的准确率。除了准确性之外,该方法还保持了很强的精度、召回率、F1和ROC-AUC,并且融合的嵌入有效地转移到多个传统的学习器(例如,Random Forest, XGBoost)。这些结果证实了该模型在计算机辅助诊断工作流程中的实际应用潜力,特别是在资源有限的临床环境中。
{"title":"Colorectal disease diagnosis with deep triple-stream fusion and attention refinement","authors":"Abdulfattah Ba Alawi ,&nbsp;Abdullah Ammar Karcioglu ,&nbsp;Ferhat Bozkurt","doi":"10.1016/j.compmedimag.2025.102669","DOIUrl":"10.1016/j.compmedimag.2025.102669","url":null,"abstract":"<div><div>Colorectal cancer constitutes a significant proportion of global cancer-related mortality, underscoring the imperative for robust and early-stage diagnostic methodologies. In this study, we propose a novel end-to-end deep learning framework that integrates multiple advanced mechanisms to enhance the classification of colorectal disease from histopathologic and endoscopic images. Our model, named <strong>TripleFusionNet</strong>, leverages a unique triple-stream architecture by combining the strengths of EfficientNetB3, ResNet50, and DenseNet121, enabling the extraction of rich, multi-level feature representations from input images. To augment discriminative feature modeling, a <em>Multi-Scale Attention Module</em> is integrated, which concurrently performs spatial and channel-wise recalibration, thereby enabling the network to emphasize diagnostically salient regions. Additionally, we incorporate a <em>Squeeze-Excite Refinement Block (SERB)</em> to selectively enhance informative channel activations while attenuating noise and redundant signals. Feature representations from the individual backbones are adaptively fused through a <em>Progressive Gated Fusion mechanism</em> that dynamically learns context-aware weighting for optimal feature integration and redundancy mitigation. We validate our approach on two colorectal benchmarks: CRCCD_V1 (14 classes) and LC25000 (binary). On CRCCD_V1, the best performance is obtained by a conventional classifier trained on our 256-D <em>TripleFusionNet</em> embeddings—SVM (RBF) reaches <strong>96.63%</strong> test accuracy with macro F1 <strong>96.62%</strong>, with the Stacking Ensemble close behind. With five-fold cross-validation, it yields comparable out-of-fold means (<strong>0.964</strong> with small standard deviations), confirming stability across partitions. End-to-end image-based baselines, including <em>TripleFusionNet</em>, are competitive but are slightly surpassed by embedding-based classifiers, highlighting the utility of the learned representation. On LC25000, our method attains <strong>100%</strong> accuracy. Beyond accuracy, the approach maintains strong precision, recall, F1, and ROC–AUC, and the fused embeddings transfer effectively to multiple conventional learners (e.g., Random Forest, XGBoost). These results confirm the potential of the model for real-world deployment in computer-aided diagnosis workflows, particularly within resource-constrained clinical settings.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102669"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145520750","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DuetMatch: Harmonizing semi-supervised brain MRI segmentation via decoupled branch optimization DuetMatch:通过解耦分支优化协调半监督脑MRI分割。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-11-11 DOI: 10.1016/j.compmedimag.2025.102666
Thanh-Huy Nguyen , Hoang-Thien Nguyen , Vi Vu , Ba-Thinh Lam , Phat Huynh , Tianyang Wang , Xingjian Li , Ulas Bagci , Min Xu
The limited availability of annotated data in medical imaging makes semi-supervised learning increasingly appealing for its ability to learn from imperfect supervision. Recently, teacher-student frameworks have gained popularity for their training benefits and robust performance. However, jointly optimizing the entire network can hinder convergence and stability, especially in challenging scenarios. To address this for medical image segmentation, we propose DuetMatch, a novel dual-branch semi-supervised framework with asynchronous optimization, where each branch optimizes either the encoder or decoder while keeping the other frozen. To improve consistency under noisy conditions, we introduce Decoupled Dropout Perturbation, enforcing regularization across branches. We also design Pairwise CutMix Cross-Guidance to enhance model diversity by exchanging pseudo-labels through augmented input pairs. To mitigate confirmation bias from noisy pseudo-labels, we propose Consistency Matching, refining labels using stable predictions from frozen teacher models. Extensive experiments on benchmark brain MRI segmentation datasets, including ISLES2022 and BraTS, show that DuetMatch consistently outperforms state-of-the-art methods, demonstrating its effectiveness and robustness across diverse semi-supervised segmentation scenarios.
医学影像中标注数据的有限可用性使得半监督学习越来越有吸引力,因为它能够从不完善的监督中学习。最近,师生框架因其培训优势和强大的性能而受到欢迎。然而,联合优化整个网络可能会影响收敛性和稳定性,特别是在具有挑战性的场景下。为了解决医学图像分割中的这个问题,我们提出了DuetMatch,这是一种具有异步优化的新型双分支半监督框架,其中每个分支优化编码器或解码器,同时保持另一个分支冻结。为了提高噪声条件下的一致性,我们引入了解耦的Dropout摄动,强制跨分支的正则化。我们还设计了Pairwise CutMix Cross-Guidance,通过增强输入对交换伪标签来增强模型的多样性。为了减轻来自噪声伪标签的确认偏差,我们提出一致性匹配,使用来自冻结教师模型的稳定预测来改进标签。在包括ISLES2022和BraTS在内的基准脑MRI分割数据集上进行的大量实验表明,DuetMatch始终优于最先进的方法,证明了其在各种半监督分割场景中的有效性和鲁棒性。
{"title":"DuetMatch: Harmonizing semi-supervised brain MRI segmentation via decoupled branch optimization","authors":"Thanh-Huy Nguyen ,&nbsp;Hoang-Thien Nguyen ,&nbsp;Vi Vu ,&nbsp;Ba-Thinh Lam ,&nbsp;Phat Huynh ,&nbsp;Tianyang Wang ,&nbsp;Xingjian Li ,&nbsp;Ulas Bagci ,&nbsp;Min Xu","doi":"10.1016/j.compmedimag.2025.102666","DOIUrl":"10.1016/j.compmedimag.2025.102666","url":null,"abstract":"<div><div>The limited availability of annotated data in medical imaging makes semi-supervised learning increasingly appealing for its ability to learn from imperfect supervision. Recently, teacher-student frameworks have gained popularity for their training benefits and robust performance. However, jointly optimizing the entire network can hinder convergence and stability, especially in challenging scenarios. To address this for medical image segmentation, we propose <em>DuetMatch</em>, a novel dual-branch semi-supervised framework with asynchronous optimization, where each branch optimizes either the encoder or decoder while keeping the other frozen. To improve consistency under noisy conditions, we introduce <strong>Decoupled Dropout Perturbation</strong>, enforcing regularization across branches. We also design <strong>Pairwise CutMix Cross-Guidance</strong> to enhance model diversity by exchanging pseudo-labels through augmented input pairs. To mitigate confirmation bias from noisy pseudo-labels, we propose <strong>Consistency Matching</strong>, refining labels using stable predictions from frozen teacher models. Extensive experiments on benchmark brain MRI segmentation datasets, including ISLES2022 and BraTS, show that DuetMatch consistently outperforms state-of-the-art methods, demonstrating its effectiveness and robustness across diverse semi-supervised segmentation scenarios.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102666"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145558034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anatomy-informed deep learning and radiomics for neurofibroma segmentation in whole-body MRI 全身MRI中神经纤维瘤分割的解剖学信息深度学习和放射组学
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-11-14 DOI: 10.1016/j.compmedimag.2025.102667
Georgii Kolokolnikov , Marie-Lena Schmalhofer , Lennart Well , Said Farschtschi , Victor-Felix Mautner , Inka Ristow , René Werner

Background and Objectives:

Neurofibromatosis type 1 (NF1) is a genetic disorder characterized by the development of multiple neurofibromas (NFs) throughout the body. Accurate segmentation of these tumors in whole-body magnetic resonance imaging (WB-MRI) is critical for quantifying tumor burden and clinical decision-making. This study aims to develop a pipeline for NF segmentation in fat-suppressed T2-weighted WB-MRI that incorporates anatomical context and radiomics to improve accuracy and specificity.

Methods:

The proposed pipeline consists of three stages: (1) anatomy segmentation using MRSegmentator and refinement with a high-risk NF zone; (2) NF segmentation using an ensemble of 3D anisotropic anatomy-informed U-Nets; and (3) tumor candidate classification using radiomic features to filter false positives. The study used 109 WB-MRI scans from 74 NF1 patients, divided into training and three test sets representing in-domain (3T), domain-shifted (1.5T), and low tumor burden scenarios. Evaluation metrics included per-scan and per-tumor Dice Similarity Coefficient (DSC), Volume Overlap Error (VOE), Absolute Relative Volume Difference (ARVD), and per-scan F1 score. Statistical significance was assessed using Wilcoxon signed-rank tests with Bonferroni correction.

Results:

On the in-domain test set, the proposed ensemble of 3D anisotropic anatomy-informed U-Nets with tumor candidate classification achieved a per-scan DSC of 0.64, outperforming 2D nnU-Net (DSC: 0.52) and 3D full-resolution nnU-Net (DSC: 0.54). Performance was maintained on the domain-shift test set (DSC: 0.51) but declined on low tumor burden cases (DSC: 0.23). Preliminary inter-reader variability analysis showed model-to-expert agreement (DSC: 0.67–0.69) comparable to inter-expert agreement (DSC: 0.69).

Conclusions:

The proposed pipeline achieves the highest performance among established methods for automated NF segmentation in WB-MRI and approaches expert-level consistency. The integration of anatomical context and radiomics enhances robustness. Nonetheless, segmentation performance decreases in low tumor burden scenarios, indicating a key area for future methodological improvements. Additionally, the limited inter-reader agreement observed among experts underscores the inherent complexity and ambiguity of the NF segmentation task.
背景和目的:1型神经纤维瘤病(NF1)是一种以全身多发性神经纤维瘤(NFs)发展为特征的遗传性疾病。在全身磁共振成像(WB-MRI)中准确分割这些肿瘤对于量化肿瘤负担和临床决策至关重要。本研究旨在开发一种在脂肪抑制的t2加权WB-MRI中进行NF分割的管道,该管道结合解剖学背景和放射组学来提高准确性和特异性。方法:该流程包括三个阶段:(1)使用MRSegmentator进行解剖分割,并使用高危NF区进行细化;(2)利用三维各向异性的U-Nets集合进行NF分割;(3)利用放射学特征对候选肿瘤进行分类,过滤假阳性。该研究使用74例NF1患者的109张WB-MRI扫描,分为训练集和三个测试集,分别代表域内(3T)、域移位(1.5T)和低肿瘤负荷情景。评估指标包括每次扫描和每个肿瘤骰子相似系数(DSC)、体积重叠误差(VOE)、绝对相对体积差(ARVD)和每次扫描F1评分。采用Wilcoxon符号秩检验和Bonferroni校正评估统计学显著性。结果:在域内测试集上,所提出的具有候选肿瘤分类的三维各向异性解剖学信息的U-Nets集合的每次扫描DSC为0.64,优于2D nnU-Net (DSC: 0.52)和3D全分辨率nnU-Net (DSC: 0.54)。在域移测试集上表现良好(DSC: 0.51),但在低肿瘤负荷情况下表现下降(DSC: 0.23)。初步的读者间变异性分析显示,模型与专家的一致性(DSC: 0.67-0.69)与专家间的一致性(DSC: 0.69)相当。结论:所提出的管道在现有的WB-MRI自动NF分割方法中达到了最高的性能,并达到了专家级的一致性。解剖学背景和放射组学的整合增强了鲁棒性。尽管如此,在低肿瘤负荷情况下,分割性能下降,这表明了未来方法改进的关键领域。此外,在专家之间观察到的有限的读者间协议强调了NF分词任务固有的复杂性和模糊性。
{"title":"Anatomy-informed deep learning and radiomics for neurofibroma segmentation in whole-body MRI","authors":"Georgii Kolokolnikov ,&nbsp;Marie-Lena Schmalhofer ,&nbsp;Lennart Well ,&nbsp;Said Farschtschi ,&nbsp;Victor-Felix Mautner ,&nbsp;Inka Ristow ,&nbsp;René Werner","doi":"10.1016/j.compmedimag.2025.102667","DOIUrl":"10.1016/j.compmedimag.2025.102667","url":null,"abstract":"<div><h3>Background and Objectives:</h3><div>Neurofibromatosis type 1 (NF1) is a genetic disorder characterized by the development of multiple neurofibromas (NFs) throughout the body. Accurate segmentation of these tumors in whole-body magnetic resonance imaging (WB-MRI) is critical for quantifying tumor burden and clinical decision-making. This study aims to develop a pipeline for NF segmentation in fat-suppressed T2-weighted WB-MRI that incorporates anatomical context and radiomics to improve accuracy and specificity.</div></div><div><h3>Methods:</h3><div>The proposed pipeline consists of three stages: (1) anatomy segmentation using MRSegmentator and refinement with a high-risk NF zone; (2) NF segmentation using an ensemble of 3D anisotropic anatomy-informed U-Nets; and (3) tumor candidate classification using radiomic features to filter false positives. The study used 109 WB-MRI scans from 74 NF1 patients, divided into training and three test sets representing in-domain (3T), domain-shifted (1.5T), and low tumor burden scenarios. Evaluation metrics included per-scan and per-tumor Dice Similarity Coefficient (DSC), Volume Overlap Error (VOE), Absolute Relative Volume Difference (ARVD), and per-scan F1 score. Statistical significance was assessed using Wilcoxon signed-rank tests with Bonferroni correction.</div></div><div><h3>Results:</h3><div>On the in-domain test set, the proposed ensemble of 3D anisotropic anatomy-informed U-Nets with tumor candidate classification achieved a per-scan DSC of 0.64, outperforming 2D nnU-Net (DSC: 0.52) and 3D full-resolution nnU-Net (DSC: 0.54). Performance was maintained on the domain-shift test set (DSC: 0.51) but declined on low tumor burden cases (DSC: 0.23). Preliminary inter-reader variability analysis showed model-to-expert agreement (DSC: 0.67–0.69) comparable to inter-expert agreement (DSC: 0.69).</div></div><div><h3>Conclusions:</h3><div>The proposed pipeline achieves the highest performance among established methods for automated NF segmentation in WB-MRI and approaches expert-level consistency. The integration of anatomical context and radiomics enhances robustness. Nonetheless, segmentation performance decreases in low tumor burden scenarios, indicating a key area for future methodological improvements. Additionally, the limited inter-reader agreement observed among experts underscores the inherent complexity and ambiguity of the NF segmentation task.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102667"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145520826","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hallucinated domain generalization network with domain-aware dynamic representation for medical image segmentation 基于域感知动态表示的幻觉域泛化网络在医学图像分割中的应用。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-11-17 DOI: 10.1016/j.compmedimag.2025.102670
Minjun Wang, Houjin Chen, Yanfeng Li, Jia Sun, Luyifu Chen, Peng Liang
Due to variations in medical image acquisition protocols, segmentation models often exhibit degraded performance when applied to unseen domains. We argue that such degradation primarily stems from overfitting to source domains and insufficient dynamic adaptability to target domains. To address this issue, we propose a hallucinated domain generalization network with domain-aware dynamic representation for medical image segmentation, which introduces a novel ”hallucination during training, dynamic representation during testing” scheme to effectively improve generalization. Specifically, we design an uncertainty-aware dynamic hallucination module that achieves adaptive transformation through Bézier curves and estimates potential domain shift by introducing the uncertainty-aware offset variable driven by channel-wise variance, generating diverse synthetic images. This approach breaks the limitations of source domain distributions while preserving original anatomical structures, effectively alleviating the model’s overfitting to the specific styles of source domains. Furthermore, we develop a domain-aware dynamic representation module that treats source domain knowledge as a foundation for understanding unknown domains. Concretely, we obtain unbiased estimates of global style prototypes through domain-wise statistical aggregation and the momentum update strategy. Then, input features are mapped to the unified source domain space through global style prototypes and similarity weights, mitigating performance degradation caused by domain shift during the testing phase. Extensive experiments on four heterogeneously distributed fundus image datasets and six multi-center prostate MRI datasets demonstrate that our approach outperforms state-of-the-art methods.
由于医学图像采集协议的变化,分割模型在应用于未知领域时往往表现出性能下降。我们认为这种退化主要源于对源域的过度拟合和对目标域的动态适应性不足。针对这一问题,我们提出了一种具有领域感知动态表示的医学图像分割幻觉域泛化网络,该网络引入了一种新颖的“训练时产生幻觉,测试时动态表示”的方案,有效地提高了医学图像的泛化效果。具体来说,我们设计了一个不确定性感知的动态幻觉模块,通过bsamizier曲线实现自适应变换,并通过引入由信道方差驱动的不确定性感知偏移变量来估计潜在的域位移,生成不同的合成图像。该方法在保留原始解剖结构的同时,打破了源域分布的限制,有效缓解了模型对源域特定样式的过拟合。此外,我们开发了一个领域感知的动态表示模块,该模块将源领域知识作为理解未知领域的基础。具体而言,我们通过领域统计聚合和动量更新策略获得了全局样式原型的无偏估计。然后,通过全局样式原型和相似度权重将输入特征映射到统一的源域空间,减轻了在测试阶段因域漂移引起的性能下降。在四个非均匀分布的眼底图像数据集和六个多中心前列腺MRI数据集上进行的大量实验表明,我们的方法优于最先进的方法。
{"title":"Hallucinated domain generalization network with domain-aware dynamic representation for medical image segmentation","authors":"Minjun Wang,&nbsp;Houjin Chen,&nbsp;Yanfeng Li,&nbsp;Jia Sun,&nbsp;Luyifu Chen,&nbsp;Peng Liang","doi":"10.1016/j.compmedimag.2025.102670","DOIUrl":"10.1016/j.compmedimag.2025.102670","url":null,"abstract":"<div><div>Due to variations in medical image acquisition protocols, segmentation models often exhibit degraded performance when applied to unseen domains. We argue that such degradation primarily stems from overfitting to source domains and insufficient dynamic adaptability to target domains. To address this issue, we propose a hallucinated domain generalization network with domain-aware dynamic representation for medical image segmentation, which introduces a novel ”hallucination during training, dynamic representation during testing” scheme to effectively improve generalization. Specifically, we design an uncertainty-aware dynamic hallucination module that achieves adaptive transformation through Bézier curves and estimates potential domain shift by introducing the uncertainty-aware offset variable driven by channel-wise variance, generating diverse synthetic images. This approach breaks the limitations of source domain distributions while preserving original anatomical structures, effectively alleviating the model’s overfitting to the specific styles of source domains. Furthermore, we develop a domain-aware dynamic representation module that treats source domain knowledge as a foundation for understanding unknown domains. Concretely, we obtain unbiased estimates of global style prototypes through domain-wise statistical aggregation and the momentum update strategy. Then, input features are mapped to the unified source domain space through global style prototypes and similarity weights, mitigating performance degradation caused by domain shift during the testing phase. Extensive experiments on four heterogeneously distributed fundus image datasets and six multi-center prostate MRI datasets demonstrate that our approach outperforms state-of-the-art methods.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"126 ","pages":"Article 102670"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145566238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computerized Medical Imaging and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1