{"title":"","authors":"","doi":"","DOIUrl":"","url":null,"abstract":"","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"128 ","pages":"Article 102715"},"PeriodicalIF":4.9,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146493490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"","authors":"","doi":"","DOIUrl":"","url":null,"abstract":"","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"128 ","pages":"Article 102721"},"PeriodicalIF":4.9,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146493493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-26DOI: 10.1016/j.compmedimag.2025.102681
Xiaodong Zhou , Huibin Wang
In the task of 3D reconstruction of X-ray coronary artery, matching vessel branches in different viewpoints is a challenging task. In this study, this task is transformed into the process of vessel branches instance segmentation and then matching branches of the same color, and an instance segmentation network (YOLO-CAVBIS) is proposed specifically for deformed and dynamic vessels. Firstly, since the left and right coronary artery branches are not easy to distinguish, a coronary artery classification dataset is produced and the left and right coronary artery arteries are classified using the YOLOv8-cls classification model, and then the classified images are fed into two parallel YOLO-CAVBIS networks for coronary artery branches instance segmentation. Finally, the branches with the same color of branches in different viewpoints are matched. The experimental results show that the accuracy of the coronary artery classification model can reach 100%, and the mAP50 of the proposed left coronary branches instance segmentation model reaches 98.4%, and the mAP50 of the proposed right coronary branches instance segmentation model reaches 99.4%. In terms of extracting deformation and dynamic vascular features, our proposed YOLO-CAVBIS network demonstrates greater specificity and superiority compared to other instance segmentation networks, and can be used as a baseline model for the task of coronary artery branches instance segmentation. Code repository: https://gitee.com/zaleman/ca_instance_segmentation, https://github.com/zaleman/ca_instance_segmentation.
{"title":"Research on X-ray coronary artery branches instance segmentation and matching task","authors":"Xiaodong Zhou , Huibin Wang","doi":"10.1016/j.compmedimag.2025.102681","DOIUrl":"10.1016/j.compmedimag.2025.102681","url":null,"abstract":"<div><div>In the task of 3D reconstruction of X-ray coronary artery, matching vessel branches in different viewpoints is a challenging task. In this study, this task is transformed into the process of vessel branches instance segmentation and then matching branches of the same color, and an instance segmentation network (YOLO-CAVBIS) is proposed specifically for deformed and dynamic vessels. Firstly, since the left and right coronary artery branches are not easy to distinguish, a coronary artery classification dataset is produced and the left and right coronary artery arteries are classified using the YOLOv8-cls classification model, and then the classified images are fed into two parallel YOLO-CAVBIS networks for coronary artery branches instance segmentation. Finally, the branches with the same color of branches in different viewpoints are matched. The experimental results show that the accuracy of the coronary artery classification model can reach 100%, and the mAP50 of the proposed left coronary branches instance segmentation model reaches 98.4%, and the mAP50 of the proposed right coronary branches instance segmentation model reaches 99.4%. In terms of extracting deformation and dynamic vascular features, our proposed YOLO-CAVBIS network demonstrates greater specificity and superiority compared to other instance segmentation networks, and can be used as a baseline model for the task of coronary artery branches instance segmentation. Code repository: <span><span>https://gitee.com/zaleman/ca_instance_segmentation</span><svg><path></path></svg></span>, <span><span>https://github.com/zaleman/ca_instance_segmentation</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"128 ","pages":"Article 102681"},"PeriodicalIF":4.9,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145928056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-26DOI: 10.1016/j.compmedimag.2025.102695
Li Shiyan, Wang Shuqin, Gu Xin, Sun Debing
In recent years, semi-supervised learning (SSL) has attracted increasing attention in medical image analysis, showing great potential in scenarios with limited annotations. However, existing consistency regularization methods suffer from several limitations: overly uniform constraints at the output layer, lack of interaction within adversarial strategies, and reliance on external sample pools for sample estimation, which together lead to insufficient use of feature-level information and unstable training. To address these challenges, this paper proposes a novel semi-supervised framework, termed Feature-level multi-scale Consistency and Adversarial Training (FCAT). A multi-scale feature-level consistency mechanism is introduced to capture hierarchical structural representations through cross-level feature fusion, enabling robust feature alignment without relying on external sample pools. To overcome the limitation of unidirectional adversarial training, a bidirectional feature perturbation strategy is designed under a teacher–student collaboration scheme, where both models generate perturbations from their own gradients and enforce mutual consistency. In addition, an intrinsic evaluation mechanism based on entropy and complementary confidence is developed to rank unlabeled samples according to their information content, guiding the training process toward informative hard samples while reducing overfitting to trivial ones. Experiments on the balanced Pneumonia Chest X-ray and NCT-CRC-HE histopathology datasets, as well as the imbalanced ISIC 2019 dermoscopic skin lesion dataset, demonstrate that our FCAT achieves competitive performance and strong generalization across diverse imaging modalities and data distributions.
{"title":"Semi-supervised medical image classification via feature-level multi-scale consistency and adversarial training","authors":"Li Shiyan, Wang Shuqin, Gu Xin, Sun Debing","doi":"10.1016/j.compmedimag.2025.102695","DOIUrl":"10.1016/j.compmedimag.2025.102695","url":null,"abstract":"<div><div>In recent years, semi-supervised learning (SSL) has attracted increasing attention in medical image analysis, showing great potential in scenarios with limited annotations. However, existing consistency regularization methods suffer from several limitations: overly uniform constraints at the output layer, lack of interaction within adversarial strategies, and reliance on external sample pools for sample estimation, which together lead to insufficient use of feature-level information and unstable training. To address these challenges, this paper proposes a novel semi-supervised framework, termed Feature-level multi-scale Consistency and Adversarial Training (FCAT). A multi-scale feature-level consistency mechanism is introduced to capture hierarchical structural representations through cross-level feature fusion, enabling robust feature alignment without relying on external sample pools. To overcome the limitation of unidirectional adversarial training, a bidirectional feature perturbation strategy is designed under a teacher–student collaboration scheme, where both models generate perturbations from their own gradients and enforce mutual consistency. In addition, an intrinsic evaluation mechanism based on entropy and complementary confidence is developed to rank unlabeled samples according to their information content, guiding the training process toward informative hard samples while reducing overfitting to trivial ones. Experiments on the balanced Pneumonia Chest X-ray and NCT-CRC-HE histopathology datasets, as well as the imbalanced ISIC 2019 dermoscopic skin lesion dataset, demonstrate that our FCAT achieves competitive performance and strong generalization across diverse imaging modalities and data distributions.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"128 ","pages":"Article 102695"},"PeriodicalIF":4.9,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145886479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-19DOI: 10.1016/j.compmedimag.2025.102690
Luohong Wu , Matthias Seibold , Nicola A. Cavalcanti , Giuseppe Loggia , Lisa Reissner , Bastian Sigrist , Jonas Hein , Lilian Calvet , Arnd Viehöfer , Philipp Fürnstahl
Background:
Bone surface reconstruction is an essential component of computer-assisted orthopedic surgery (CAOS), forming the foundation for both preoperative planning and intraoperative guidance. Compared to traditional imaging modalities such as computed tomography (CT) and magnetic resonance imaging (MRI), ultrasound, an emerging CAOS technology, provides a radiation-free, cost-effective, and portable alternative. While ultrasound offers new opportunities in CAOS, technical shortcomings continue to hinder its translation into surgery. In particular, due to the inherent limitations of ultrasound imaging, B-mode ultrasound typically captures only partial bone surfaces. The inter- and intra-operator variability in ultrasound scanning further increases the complexity of the data. Existing reconstruction methods struggle with such challenging data, leading to increased reconstruction errors and artifacts, such as holes and inflated structures. Effective techniques for accurately reconstructing open bone surfaces from real-world 3D ultrasound volumes remain lacking.
Methods:
We propose UltraBoneUDF, a self-supervised framework specifically designed for reconstructing open bone surfaces from ultrasound data. It learns unsigned distance functions (UDFs) from 3D ultrasound data. In addition, we present a novel loss function based on local tangent plane optimization that substantially improves surface reconstruction quality. UltraBoneUDF and competing models are benchmarked on three open-source datasets and further evaluated through ablation studies.
Results:
Qualitative results demonstrate the limitations of the state-of-the-art methods. Quantitatively, UltraBoneUDF achieves comparable or lower bi-directional Chamfer distance across three datasets with fewer parameters: 1.60 mm on the UltraBones100k dataset ( improvement), 0.21 mm on the OpenBoneCT dataset, and 0.18 mm on the ClosedBoneCT dataset.
Conclusion:
UltraBoneUDF represents a promising solution for open bone surface reconstruction from 3D ultrasound volumes, with the potential to advance downstream applications in CAOS.
{"title":"UltraBoneUDF: Self-supervised bone surface reconstruction from ultrasound based on neural unsigned distance functions","authors":"Luohong Wu , Matthias Seibold , Nicola A. Cavalcanti , Giuseppe Loggia , Lisa Reissner , Bastian Sigrist , Jonas Hein , Lilian Calvet , Arnd Viehöfer , Philipp Fürnstahl","doi":"10.1016/j.compmedimag.2025.102690","DOIUrl":"10.1016/j.compmedimag.2025.102690","url":null,"abstract":"<div><h3>Background:</h3><div>Bone surface reconstruction is an essential component of computer-assisted orthopedic surgery (CAOS), forming the foundation for both preoperative planning and intraoperative guidance. Compared to traditional imaging modalities such as computed tomography (CT) and magnetic resonance imaging (MRI), ultrasound, an emerging CAOS technology, provides a radiation-free, cost-effective, and portable alternative. While ultrasound offers new opportunities in CAOS, technical shortcomings continue to hinder its translation into surgery. In particular, due to the inherent limitations of ultrasound imaging, B-mode ultrasound typically captures only partial bone surfaces. The inter- and intra-operator variability in ultrasound scanning further increases the complexity of the data. Existing reconstruction methods struggle with such challenging data, leading to increased reconstruction errors and artifacts, such as holes and inflated structures. Effective techniques for accurately reconstructing open bone surfaces from real-world 3D ultrasound volumes remain lacking.</div></div><div><h3>Methods:</h3><div>We propose UltraBoneUDF, a self-supervised framework specifically designed for reconstructing open bone surfaces from ultrasound data. It learns unsigned distance functions (UDFs) from 3D ultrasound data. In addition, we present a novel loss function based on local tangent plane optimization that substantially improves surface reconstruction quality. UltraBoneUDF and competing models are benchmarked on three open-source datasets and further evaluated through ablation studies.</div></div><div><h3>Results:</h3><div>Qualitative results demonstrate the limitations of the state-of-the-art methods. Quantitatively, UltraBoneUDF achieves comparable or lower bi-directional Chamfer distance across three datasets with fewer parameters: 1.60 mm on the UltraBones100k dataset (<span><math><mrow><mo>≈</mo><mn>25</mn><mo>.</mo><mn>5</mn><mtext>%</mtext></mrow></math></span> improvement), 0.21 mm on the OpenBoneCT dataset, and 0.18 mm on the ClosedBoneCT dataset.</div></div><div><h3>Conclusion:</h3><div>UltraBoneUDF represents a promising solution for open bone surface reconstruction from 3D ultrasound volumes, with the potential to advance downstream applications in CAOS.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102690"},"PeriodicalIF":4.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145840092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-19DOI: 10.1016/j.compmedimag.2025.102689
Yuheng Yang , Kun You , Haoyang He , Yuehua Zhang , Xue Feng , Fei Lu , Luping Fang , Yunling Wang , Qing Pan
Deep learning methods have been widely used in medical imaging, including rib segmentation. Nevertheless, their dependence on large annotated datasets poses a significant challenge, as expert-annotated rib Computed Tomography (CT) scans are notably scarce. An additional complication arises from domain shift, which often limits the direct applicability of models trained on public datasets to specific clinical tasks, thus requiring further resource-intensive annotations on target domains for adaptation. Although semi-supervised methods have been developed to mitigate annotation costs, the prevailing strategies largely remain at the sample level. This results in unavoidable redundancy within each annotated sample, making the process of labeling an entire CT scan exceedingly tedious and costly. To address these issues, we propose a semi-supervised approach named the Entropy-Guided Partial Annotation (EGPA) method for rib segmentation. This method actively identifies the most informative regions in images for annotation based on entropy metrics, thereby substantially reducing the workload for experts during both model training and cross-domain adaptation. By integrating contrastive learning, active learning, and self-training strategies, EGPA not only significantly saves annotation cost and time when training from scratch but also effectively addresses the challenges of migrating from source to target domains. On the public RibSegV2 dataset (source domain) and a private chest CT rib segmentation dataset (target domain), EGPA achieved Dice scores of 89.5 and 90.7, respectively, nearly matching the performance of fully supervised models (89.9 and 91.2) with only 19 % and 18 % of the full annotation workload. This remarkable reduction in annotation effort shortens the development timeline for reliable segmentation tools and enhances their clinical feasibility. By simplifying the creation of high-quality annotated datasets, our approach facilitates the broad deployment of rib analysis tools in varied clinical settings, promoting standardized and efficient diagnostic practices.
{"title":"Entropy-guided partial annotation for cross-domain rib segmentation","authors":"Yuheng Yang , Kun You , Haoyang He , Yuehua Zhang , Xue Feng , Fei Lu , Luping Fang , Yunling Wang , Qing Pan","doi":"10.1016/j.compmedimag.2025.102689","DOIUrl":"10.1016/j.compmedimag.2025.102689","url":null,"abstract":"<div><div>Deep learning methods have been widely used in medical imaging, including rib segmentation. Nevertheless, their dependence on large annotated datasets poses a significant challenge, as expert-annotated rib Computed Tomography (CT) scans are notably scarce. An additional complication arises from domain shift, which often limits the direct applicability of models trained on public datasets to specific clinical tasks, thus requiring further resource-intensive annotations on target domains for adaptation. Although semi-supervised methods have been developed to mitigate annotation costs, the prevailing strategies largely remain at the sample level. This results in unavoidable redundancy within each annotated sample, making the process of labeling an entire CT scan exceedingly tedious and costly. To address these issues, we propose a semi-supervised approach named the Entropy-Guided Partial Annotation (EGPA) method for rib segmentation. This method actively identifies the most informative regions in images for annotation based on entropy metrics, thereby substantially reducing the workload for experts during both model training and cross-domain adaptation. By integrating contrastive learning, active learning, and self-training strategies, EGPA not only significantly saves annotation cost and time when training from scratch but also effectively addresses the challenges of migrating from source to target domains. On the public RibSegV2 dataset (source domain) and a private chest CT rib segmentation dataset (target domain), EGPA achieved Dice scores of 89.5 and 90.7, respectively, nearly matching the performance of fully supervised models (89.9 and 91.2) with only 19 % and 18 % of the full annotation workload. This remarkable reduction in annotation effort shortens the development timeline for reliable segmentation tools and enhances their clinical feasibility. By simplifying the creation of high-quality annotated datasets, our approach facilitates the broad deployment of rib analysis tools in varied clinical settings, promoting standardized and efficient diagnostic practices.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102689"},"PeriodicalIF":4.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145835241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-19DOI: 10.1016/j.compmedimag.2025.102692
Qiongmin Zhang, Siyi Yu, Yin Shi, Xiaowei Tan
Alzheimer’s Disease (AD) is a progressive neurodegenerative disorder characterized by cognitive decline and memory loss. Magnetic Resonance Imaging (MRI) plays a key role in detecting AD in Computer-aided Diagnosis (CAD) systems. However, variations in MRI scanners and imaging protocols introduce domain shifts, which significantly degrade model performance. Additionally, CAD models may misdiagnose unfamiliar neurodegenerative diseases not represented during training. In these complex and diverse clinical scenarios, employing closed set domain adaptation methods to achieve accurate diagnosis of AD presents substantial challenges. We propose a Hyperspherical Weighted Adversarial Learning-based Open Set Domain Adaptation (HWAL-OSDA) method for AD diagnosis. We introduce a voxel-based 3D feature extraction and fusion module to effectively capture and integrate MRI spatial features and employ a Multi-scale and Dual Attention Aggregation block to focus on disease-sensitive regions. To overcome the dispersion of feature distributions in high-dimensional space, hyperspherical variational auto-encoder module is incorporated to improve the learning of latent feature representations on a hypersphere. Furthermore, the spherical angular distance-based triplet loss and margin-based loss in the cross-domain alignment and separation module enhance the separability of known classes and establish a clear decision boundary between known and unknown classes. To improve the positive transfer of known samples and reduce the negative transfer of unknown samples, we design a weighted adversarial domain adaptation module that utilizes a dynamic instance-level weighting scheme, combining the Weibull distribution with entropy. Experiments on the ADNI and PPMI datasets show that HWAL-OSDA achieves an average accuracy of 94.2%, 83.68%, and 77.83% across three-way classification tasks (AD vs. CN vs. Unk, MCI vs. CN vs. Unk, and AD vs. MCI vs. Unk tasks), outperforming traditional and state-of-the-art OSDA methods. This approach offers a practical reference for CAD of AD and other neurodegenerative diseases in open clinical settings.
阿尔茨海默病(AD)是一种以认知能力下降和记忆丧失为特征的进行性神经退行性疾病。在计算机辅助诊断(CAD)系统中,磁共振成像(MRI)在检测AD方面起着关键作用。然而,MRI扫描仪和成像协议的变化引入了域移位,这大大降低了模型的性能。此外,CAD模型可能误诊训练中未出现的不熟悉的神经退行性疾病。在这些复杂多样的临床场景中,采用闭集域自适应方法实现AD的准确诊断提出了很大的挑战。提出了一种基于超球面加权对抗学习的开集域自适应(hwalosda)的AD诊断方法。引入基于体素的三维特征提取与融合模块,有效捕获和整合MRI空间特征,采用多尺度双注意力聚合块对疾病敏感区域进行聚焦。为了克服特征分布在高维空间中的分散性,引入了超球变分自编码器模块,提高了对超球上潜在特征表示的学习。此外,跨域对准与分离模块中基于球面角距离的三重态损失和基于边缘的损失增强了已知类的可分性,并在已知和未知类之间建立了明确的决策边界。为了提高已知样本的正迁移和减少未知样本的负迁移,我们设计了一个加权的对抗域自适应模块,该模块采用动态实例级加权方案,将威布尔分布与熵相结合。在ADNI和PPMI数据集上的实验表明,hwo -OSDA在三向分类任务(AD vs. CN vs. Unk, MCI vs. CN vs. Unk, AD vs. MCI vs. Unk)上的平均准确率分别为94.2%、83.68%和77.83%,优于传统和最先进的OSDA方法。该方法可为开放式临床环境下AD及其他神经退行性疾病的CAD提供实用参考。
{"title":"Improving Alzheimer’s disease diagnosis by hyperspherical weighted adversarial learning in open set domain adaptation","authors":"Qiongmin Zhang, Siyi Yu, Yin Shi, Xiaowei Tan","doi":"10.1016/j.compmedimag.2025.102692","DOIUrl":"10.1016/j.compmedimag.2025.102692","url":null,"abstract":"<div><div>Alzheimer’s Disease (AD) is a progressive neurodegenerative disorder characterized by cognitive decline and memory loss. Magnetic Resonance Imaging (MRI) plays a key role in detecting AD in Computer-aided Diagnosis (CAD) systems. However, variations in MRI scanners and imaging protocols introduce domain shifts, which significantly degrade model performance. Additionally, CAD models may misdiagnose unfamiliar neurodegenerative diseases not represented during training. In these complex and diverse clinical scenarios, employing closed set domain adaptation methods to achieve accurate diagnosis of AD presents substantial challenges. We propose a Hyperspherical Weighted Adversarial Learning-based Open Set Domain Adaptation (HWAL-OSDA) method for AD diagnosis. We introduce a voxel-based 3D feature extraction and fusion module to effectively capture and integrate MRI spatial features and employ a Multi-scale and Dual Attention Aggregation block to focus on disease-sensitive regions. To overcome the dispersion of feature distributions in high-dimensional space, hyperspherical variational auto-encoder module is incorporated to improve the learning of latent feature representations on a hypersphere. Furthermore, the spherical angular distance-based triplet loss and margin-based loss in the cross-domain alignment and separation module enhance the separability of known classes and establish a clear decision boundary between known and unknown classes. To improve the positive transfer of known samples and reduce the negative transfer of unknown samples, we design a weighted adversarial domain adaptation module that utilizes a dynamic instance-level weighting scheme, combining the Weibull distribution with entropy. Experiments on the ADNI and PPMI datasets show that HWAL-OSDA achieves an average accuracy of 94.2%, 83.68%, and 77.83% across three-way classification tasks (AD vs. CN vs. Unk, MCI vs. CN vs. Unk, and AD vs. MCI vs. Unk tasks), outperforming traditional and state-of-the-art OSDA methods. This approach offers a practical reference for CAD of AD and other neurodegenerative diseases in open clinical settings.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102692"},"PeriodicalIF":4.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-17DOI: 10.1016/j.compmedimag.2025.102685
Wei Chen , Fan Wu , Yan Guo , Weiqing Zhuang , Hua Chen , Jie Feng , Jianfeng Wu
Lumbar spinal stenosis (LSS) represents a significant global health burden, and its diagnosis from Magnetic Resonance Imaging (MRI) is often subject to inter-observer variability and time-consuming interpretation. While deep learning (DL) models offer a promising solution, they are frequently constrained by the scarcity of annotated medical data, high computational demands, and challenges in representing subtle pathological features. To address these limitations, we propose KWC-YOLO, a novel and efficient object detection framework for the automated detection and classification of lumbar central canal stenosis (LCCS) severity according to the Schizas grading criteria. Our model enhances the YOLOv11n architecture through three core innovations: (1) the integration of KernelWarehouse (KWConv), a parameter-efficient dynamic convolution mechanism that improves the feature adaptability of the detection head; (2) the introduction of a FasterGATE activation unit in the backbone to enhance non-linear representation and accelerate convergence; and (3) the implementation of a lightweight Slim-Neck structure, which optimizes the trade-off between feature fusion quality and computational cost. On a clinical lumbar spine MRI dataset, KWC-YOLO demonstrates superior performance, achieving a mean Average Precision at an IoU of 0.5 () of 86.7% and an of 63.0%. This represents a substantial improvement over the YOLOv11n baseline by 9.2 and 9.3 percentage points in and respectively, while simultaneously reducing the computational load by 36.5% to 4.0 GFLOPs. Conclusively, KWC-YOLO establishes a new benchmark for automated LCCS grading. Its compelling balance of high accuracy and computational efficiency holds the potential to alleviate the interpretative burden on radiologists, enhance reporting accuracy, and streamline clinical decision-making, ultimately leading to improved patient outcomes.
{"title":"KWC-YOLO: An efficient YOLO architecture for lumbar spinal stenosis grading through dynamic convolution and spatially-aware gating","authors":"Wei Chen , Fan Wu , Yan Guo , Weiqing Zhuang , Hua Chen , Jie Feng , Jianfeng Wu","doi":"10.1016/j.compmedimag.2025.102685","DOIUrl":"10.1016/j.compmedimag.2025.102685","url":null,"abstract":"<div><div>Lumbar spinal stenosis (LSS) represents a significant global health burden, and its diagnosis from Magnetic Resonance Imaging (MRI) is often subject to inter-observer variability and time-consuming interpretation. While deep learning (DL) models offer a promising solution, they are frequently constrained by the scarcity of annotated medical data, high computational demands, and challenges in representing subtle pathological features. To address these limitations, we propose KWC-YOLO, a novel and efficient object detection framework for the automated detection and classification of lumbar central canal stenosis (LCCS) severity according to the Schizas grading criteria. Our model enhances the YOLOv11n architecture through three core innovations: (1) the integration of KernelWarehouse (KWConv), a parameter-efficient dynamic convolution mechanism that improves the feature adaptability of the detection head; (2) the introduction of a FasterGATE activation unit in the backbone to enhance non-linear representation and accelerate convergence; and (3) the implementation of a lightweight Slim-Neck structure, which optimizes the trade-off between feature fusion quality and computational cost. On a clinical lumbar spine MRI dataset, KWC-YOLO demonstrates superior performance, achieving a mean Average Precision at an IoU of 0.5 (<span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>50</mn></mrow></msub></mrow></math></span>) of 86.7% and an <span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>95</mn></mrow></msub></mrow></math></span> of 63.0%. This represents a substantial improvement over the YOLOv11n baseline by 9.2 and 9.3 percentage points in <span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>50</mn></mrow></msub></mrow></math></span> and <span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>95</mn></mrow></msub></mrow></math></span> respectively, while simultaneously reducing the computational load by 36.5% to 4.0 GFLOPs. Conclusively, KWC-YOLO establishes a new benchmark for automated LCCS grading. Its compelling balance of high accuracy and computational efficiency holds the potential to alleviate the interpretative burden on radiologists, enhance reporting accuracy, and streamline clinical decision-making, ultimately leading to improved patient outcomes.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102685"},"PeriodicalIF":4.9,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145840029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-08DOI: 10.1016/j.compmedimag.2025.102686
Yueyu Huang , Lu Qiang , Wenyu Xing , Tao Jiang , Yucheng He , Yuan Liu , Gaobo Zhang , Jingxian Wang , Xiaojun Song , Yifang Li
Generating a CT image from an existing but unpaired MR image has recently become a promising objective in the radiotherapy treatment planning (RTP), whereas the unsatisfactory structure and detail accuracy of the outcome is still a challenge for its clinical application. To address this issue, this work proposed an unsupervised model called structure-adaptive normalization wavelet (SAN-Wavelet) cycleGAN for unpaired MR-to-CT image synthesis. This method not only developed a module called SAN to ensure the anatomically structural integrity, but also introduced a neighborhood detail loss (ND loss) for detail consistency penalization between different modalities. Furthermore, a high-frequency discriminator and wavelet-trans skip connection were designed to keep with the high-frequency tissue detail. Experimental findings on abdominal area demonstrated the effectiveness of SAN-Wavelet cycleGAN for unpaired MR-to-CT synthesis, with mean squared error (MSE) of 66.38, root mean squared error (RMSE) of 8.07, peak signal-to-noise ratio (PSNR) of 25.944 dB, structural similarity index (SSlM) of 0.895 and mixture perceptual similarity index (MPSIM) of 0.723. Compared to other unsupervised approaches (i.e., cycleGAN, gc-, cc-, and sc-cycleGAN), the performances of SAN-cycleGAN improved by 5–15 % in terms of the metrics above. Moreover, the dosimetric distributions of the synthesized CT and real CT in planning target volume (PTV-45Gy) and organ at risk area (bowel) were statistically consistent (Mann-Whitney U test, P > 0.05) with a mean relative error being 0.09 % and 0.24 %, showcasing a promising applicability for clinical application.
{"title":"SAN-Wavelet cycleGAN for anatomical structure reinforcement and tissue detail preservation in abdominal CT synthesis","authors":"Yueyu Huang , Lu Qiang , Wenyu Xing , Tao Jiang , Yucheng He , Yuan Liu , Gaobo Zhang , Jingxian Wang , Xiaojun Song , Yifang Li","doi":"10.1016/j.compmedimag.2025.102686","DOIUrl":"10.1016/j.compmedimag.2025.102686","url":null,"abstract":"<div><div>Generating a CT image from an existing but unpaired MR image has recently become a promising objective in the radiotherapy treatment planning (RTP), whereas the unsatisfactory structure and detail accuracy of the outcome is still a challenge for its clinical application. To address this issue, this work proposed an unsupervised model called structure-adaptive normalization wavelet (SAN-Wavelet) cycleGAN for unpaired MR-to-CT image synthesis. This method not only developed a module called SAN to ensure the anatomically structural integrity, but also introduced a neighborhood detail loss (ND loss) for detail consistency penalization between different modalities. Furthermore, a high-frequency discriminator and wavelet-trans skip connection were designed to keep with the high-frequency tissue detail. Experimental findings on abdominal area demonstrated the effectiveness of SAN-Wavelet cycleGAN for unpaired MR-to-CT synthesis, with mean squared error (MSE) of 66.38, root mean squared error (RMSE) of 8.07, peak signal-to-noise ratio (PSNR) of 25.944 dB, structural similarity index (SSlM) of 0.895 and mixture perceptual similarity index (MPSIM) of 0.723. Compared to other unsupervised approaches (i.e., cycleGAN, gc-, cc-, and sc-cycleGAN), the performances of SAN-cycleGAN improved by 5–15 % in terms of the metrics above. Moreover, the dosimetric distributions of the synthesized CT and real CT in planning target volume (PTV-45Gy) and organ at risk area (bowel) were statistically consistent (Mann-Whitney U test, P > 0.05) with a mean relative error being 0.09 % and 0.24 %, showcasing a promising applicability for clinical application.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102686"},"PeriodicalIF":4.9,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145758295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-08DOI: 10.1016/j.compmedimag.2025.102687
Younes Akbari , Faseela Abdullakutty , Somaya Al-Maadeed , Rafif Al Saady , Ahmed Bouridane , Rifat Hamoudi
<div><div>Multimodal medical imaging combining conventional imaging modalities such as mammography, ultrasound, and histopathology has shown significant promise for improving breast cancer detection accuracy. However, clinical implementation faces substantial challenges due to incomplete patient-matched multimodal datasets and resource constraints. Traditional approaches require complete imaging workups from individual patients, limiting their practical applicability. This study investigates whether cross-patient multimodal fusion combining imaging modalities from different patients, can provide additional diagnostic information beyond single-modality approaches. We hypothesize that leveraging complementary information from heterogeneous patient populations enhances cancer detection performance, even when modalities originate from separate individuals. We developed a novel virtual patient framework that systematically combines imaging modalities across different patients based on quality-driven selection strategies. Two training paradigms were evaluated: Fixed scenario with 1:1:1 cross-patient combinations (<span><math><mo>∼</mo></math></span>250 virtual patients), and Combinatorial scenario with systematic companion selection (<span><math><mo>∼</mo></math></span>20,000 virtual patients). Multiple fusion architectures (concatenation, attention, and averaging) were assessed, and we designed a novel co-attention mechanism that enables sophisticated cross-modal interaction through learned attention weights. These fusion networks were evaluated using histopathology (BCSS), mammography, and ultrasound (BUSI) datasets. External validation using the ICIAR2018 BACH Challenge dataset as an alternative histopathology source demonstrated the generalizability of our approach, achieving promising accuracy despite differences in staining protocols and acquisition procedures across institutions. All models were evaluated on consistent fixed test sets to ensure fair comparison. This dataset is well-suited for multiple breast cancer analysis tasks, including detection, segmentation, and Explainable Artificial Intelligence (XAI) applications. Cross-patient multimodal fusion demonstrated significant improvements over single-modality approaches. The best single modality achieved 75.36% accuracy (mammography), while the optimal fusion combination (histopathology-mammography) reached 97.10% accuracy, representing a 21.74 percentage point improvement. Comprehensive quantitative validation through silhouette analysis (score: 0.894) confirms that the observed performance improvements reflect genuine feature space structure rather than visualization artifacts. Cross-patient multimodal fusion demonstrates significant potential for enhancing breast cancer detection, particularly addressing real-world scenarios where complete patient-matched multimodal data is unavailable. This approach represents a paradigm shift toward leveraging heterogeneous information sources for impro
{"title":"A novel virtual patient approach for cross-patient multimodal fusion in enhanced breast cancer detection","authors":"Younes Akbari , Faseela Abdullakutty , Somaya Al-Maadeed , Rafif Al Saady , Ahmed Bouridane , Rifat Hamoudi","doi":"10.1016/j.compmedimag.2025.102687","DOIUrl":"10.1016/j.compmedimag.2025.102687","url":null,"abstract":"<div><div>Multimodal medical imaging combining conventional imaging modalities such as mammography, ultrasound, and histopathology has shown significant promise for improving breast cancer detection accuracy. However, clinical implementation faces substantial challenges due to incomplete patient-matched multimodal datasets and resource constraints. Traditional approaches require complete imaging workups from individual patients, limiting their practical applicability. This study investigates whether cross-patient multimodal fusion combining imaging modalities from different patients, can provide additional diagnostic information beyond single-modality approaches. We hypothesize that leveraging complementary information from heterogeneous patient populations enhances cancer detection performance, even when modalities originate from separate individuals. We developed a novel virtual patient framework that systematically combines imaging modalities across different patients based on quality-driven selection strategies. Two training paradigms were evaluated: Fixed scenario with 1:1:1 cross-patient combinations (<span><math><mo>∼</mo></math></span>250 virtual patients), and Combinatorial scenario with systematic companion selection (<span><math><mo>∼</mo></math></span>20,000 virtual patients). Multiple fusion architectures (concatenation, attention, and averaging) were assessed, and we designed a novel co-attention mechanism that enables sophisticated cross-modal interaction through learned attention weights. These fusion networks were evaluated using histopathology (BCSS), mammography, and ultrasound (BUSI) datasets. External validation using the ICIAR2018 BACH Challenge dataset as an alternative histopathology source demonstrated the generalizability of our approach, achieving promising accuracy despite differences in staining protocols and acquisition procedures across institutions. All models were evaluated on consistent fixed test sets to ensure fair comparison. This dataset is well-suited for multiple breast cancer analysis tasks, including detection, segmentation, and Explainable Artificial Intelligence (XAI) applications. Cross-patient multimodal fusion demonstrated significant improvements over single-modality approaches. The best single modality achieved 75.36% accuracy (mammography), while the optimal fusion combination (histopathology-mammography) reached 97.10% accuracy, representing a 21.74 percentage point improvement. Comprehensive quantitative validation through silhouette analysis (score: 0.894) confirms that the observed performance improvements reflect genuine feature space structure rather than visualization artifacts. Cross-patient multimodal fusion demonstrates significant potential for enhancing breast cancer detection, particularly addressing real-world scenarios where complete patient-matched multimodal data is unavailable. This approach represents a paradigm shift toward leveraging heterogeneous information sources for impro","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102687"},"PeriodicalIF":4.9,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145716425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}