首页 > 最新文献

Computerized Medical Imaging and Graphics最新文献

英文 中文
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-01
{"title":"","authors":"","doi":"","DOIUrl":"","url":null,"abstract":"","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"128 ","pages":"Article 102715"},"PeriodicalIF":4.9,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146493490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-01
{"title":"","authors":"","doi":"","DOIUrl":"","url":null,"abstract":"","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"128 ","pages":"Article 102721"},"PeriodicalIF":4.9,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146493493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on X-ray coronary artery branches instance segmentation and matching task x射线冠状动脉分支实例分割与匹配任务研究
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-26 DOI: 10.1016/j.compmedimag.2025.102681
Xiaodong Zhou , Huibin Wang
In the task of 3D reconstruction of X-ray coronary artery, matching vessel branches in different viewpoints is a challenging task. In this study, this task is transformed into the process of vessel branches instance segmentation and then matching branches of the same color, and an instance segmentation network (YOLO-CAVBIS) is proposed specifically for deformed and dynamic vessels. Firstly, since the left and right coronary artery branches are not easy to distinguish, a coronary artery classification dataset is produced and the left and right coronary artery arteries are classified using the YOLOv8-cls classification model, and then the classified images are fed into two parallel YOLO-CAVBIS networks for coronary artery branches instance segmentation. Finally, the branches with the same color of branches in different viewpoints are matched. The experimental results show that the accuracy of the coronary artery classification model can reach 100%, and the mAP50 of the proposed left coronary branches instance segmentation model reaches 98.4%, and the mAP50 of the proposed right coronary branches instance segmentation model reaches 99.4%. In terms of extracting deformation and dynamic vascular features, our proposed YOLO-CAVBIS network demonstrates greater specificity and superiority compared to other instance segmentation networks, and can be used as a baseline model for the task of coronary artery branches instance segmentation. Code repository: https://gitee.com/zaleman/ca_instance_segmentation, https://github.com/zaleman/ca_instance_segmentation.
在x线冠状动脉三维重建任务中,不同视点的血管分支匹配是一项具有挑战性的任务。在本研究中,将该任务转化为血管分支实例分割和相同颜色分支匹配的过程,并提出了针对变形血管和动态血管的实例分割网络(YOLO-CAVBIS)。首先,针对左右冠状动脉分支不易区分的问题,建立冠状动脉分类数据集,利用YOLOv8-cls分类模型对左右冠状动脉进行分类,然后将分类后的图像送入两个并行的yolov8 - cavbis网络进行冠状动脉分支实例分割。最后,对不同视点分支颜色相同的分支进行匹配。实验结果表明,冠状动脉分类模型的准确率可以达到100%,所提出的左冠状动脉分支实例分割模型的mAP50达到98.4%,所提出的右冠状动脉分支实例分割模型的mAP50达到99.4%。在提取血管形变和血管动态特征方面,与其他实例分割网络相比,我们提出的YOLO-CAVBIS网络具有更大的特异性和优越性,可以作为冠状动脉分支实例分割任务的基线模型。代码存储库:https://gitee.com/zaleman/ca_instance_segmentation, https://github.com/zaleman/ca_instance_segmentation。
{"title":"Research on X-ray coronary artery branches instance segmentation and matching task","authors":"Xiaodong Zhou ,&nbsp;Huibin Wang","doi":"10.1016/j.compmedimag.2025.102681","DOIUrl":"10.1016/j.compmedimag.2025.102681","url":null,"abstract":"<div><div>In the task of 3D reconstruction of X-ray coronary artery, matching vessel branches in different viewpoints is a challenging task. In this study, this task is transformed into the process of vessel branches instance segmentation and then matching branches of the same color, and an instance segmentation network (YOLO-CAVBIS) is proposed specifically for deformed and dynamic vessels. Firstly, since the left and right coronary artery branches are not easy to distinguish, a coronary artery classification dataset is produced and the left and right coronary artery arteries are classified using the YOLOv8-cls classification model, and then the classified images are fed into two parallel YOLO-CAVBIS networks for coronary artery branches instance segmentation. Finally, the branches with the same color of branches in different viewpoints are matched. The experimental results show that the accuracy of the coronary artery classification model can reach 100%, and the mAP50 of the proposed left coronary branches instance segmentation model reaches 98.4%, and the mAP50 of the proposed right coronary branches instance segmentation model reaches 99.4%. In terms of extracting deformation and dynamic vascular features, our proposed YOLO-CAVBIS network demonstrates greater specificity and superiority compared to other instance segmentation networks, and can be used as a baseline model for the task of coronary artery branches instance segmentation. Code repository: <span><span>https://gitee.com/zaleman/ca_instance_segmentation</span><svg><path></path></svg></span>, <span><span>https://github.com/zaleman/ca_instance_segmentation</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"128 ","pages":"Article 102681"},"PeriodicalIF":4.9,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145928056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-supervised medical image classification via feature-level multi-scale consistency and adversarial training 基于特征级多尺度一致性和对抗训练的半监督医学图像分类
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-26 DOI: 10.1016/j.compmedimag.2025.102695
Li Shiyan, Wang Shuqin, Gu Xin, Sun Debing
In recent years, semi-supervised learning (SSL) has attracted increasing attention in medical image analysis, showing great potential in scenarios with limited annotations. However, existing consistency regularization methods suffer from several limitations: overly uniform constraints at the output layer, lack of interaction within adversarial strategies, and reliance on external sample pools for sample estimation, which together lead to insufficient use of feature-level information and unstable training. To address these challenges, this paper proposes a novel semi-supervised framework, termed Feature-level multi-scale Consistency and Adversarial Training (FCAT). A multi-scale feature-level consistency mechanism is introduced to capture hierarchical structural representations through cross-level feature fusion, enabling robust feature alignment without relying on external sample pools. To overcome the limitation of unidirectional adversarial training, a bidirectional feature perturbation strategy is designed under a teacher–student collaboration scheme, where both models generate perturbations from their own gradients and enforce mutual consistency. In addition, an intrinsic evaluation mechanism based on entropy and complementary confidence is developed to rank unlabeled samples according to their information content, guiding the training process toward informative hard samples while reducing overfitting to trivial ones. Experiments on the balanced Pneumonia Chest X-ray and NCT-CRC-HE histopathology datasets, as well as the imbalanced ISIC 2019 dermoscopic skin lesion dataset, demonstrate that our FCAT achieves competitive performance and strong generalization across diverse imaging modalities and data distributions.
近年来,半监督学习(semi-supervised learning, SSL)在医学图像分析中受到越来越多的关注,在标注有限的场景中显示出巨大的潜力。然而,现有的一致性正则化方法存在一些局限性:输出层过于统一的约束,对抗策略之间缺乏交互,以及依赖外部样本池进行样本估计,这些都导致特征级信息的使用不足和训练不稳定。为了解决这些挑战,本文提出了一种新的半监督框架,称为特征级多尺度一致性和对抗训练(FCAT)。引入了一种多尺度特征级一致性机制,通过跨级别特征融合捕获分层结构表示,实现了不依赖外部样本池的鲁棒特征对齐。为了克服单向对抗训练的局限性,在师生协作方案下设计了双向特征摄动策略,两个模型从各自的梯度产生摄动,并实现相互一致性。此外,开发了一种基于熵和互补置信度的内在评价机制,根据未标记样本的信息内容对其进行排序,引导训练过程向信息丰富的硬样本发展,同时减少对平凡样本的过拟合。在平衡的肺炎胸片和NCT-CRC-HE组织病理学数据集以及不平衡的ISIC 2019皮肤镜皮肤病变数据集上进行的实验表明,我们的FCAT在不同的成像方式和数据分布中具有竞争力的性能和很强的泛化性。
{"title":"Semi-supervised medical image classification via feature-level multi-scale consistency and adversarial training","authors":"Li Shiyan,&nbsp;Wang Shuqin,&nbsp;Gu Xin,&nbsp;Sun Debing","doi":"10.1016/j.compmedimag.2025.102695","DOIUrl":"10.1016/j.compmedimag.2025.102695","url":null,"abstract":"<div><div>In recent years, semi-supervised learning (SSL) has attracted increasing attention in medical image analysis, showing great potential in scenarios with limited annotations. However, existing consistency regularization methods suffer from several limitations: overly uniform constraints at the output layer, lack of interaction within adversarial strategies, and reliance on external sample pools for sample estimation, which together lead to insufficient use of feature-level information and unstable training. To address these challenges, this paper proposes a novel semi-supervised framework, termed Feature-level multi-scale Consistency and Adversarial Training (FCAT). A multi-scale feature-level consistency mechanism is introduced to capture hierarchical structural representations through cross-level feature fusion, enabling robust feature alignment without relying on external sample pools. To overcome the limitation of unidirectional adversarial training, a bidirectional feature perturbation strategy is designed under a teacher–student collaboration scheme, where both models generate perturbations from their own gradients and enforce mutual consistency. In addition, an intrinsic evaluation mechanism based on entropy and complementary confidence is developed to rank unlabeled samples according to their information content, guiding the training process toward informative hard samples while reducing overfitting to trivial ones. Experiments on the balanced Pneumonia Chest X-ray and NCT-CRC-HE histopathology datasets, as well as the imbalanced ISIC 2019 dermoscopic skin lesion dataset, demonstrate that our FCAT achieves competitive performance and strong generalization across diverse imaging modalities and data distributions.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"128 ","pages":"Article 102695"},"PeriodicalIF":4.9,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145886479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UltraBoneUDF: Self-supervised bone surface reconstruction from ultrasound based on neural unsigned distance functions UltraBoneUDF:基于神经无符号距离函数的超声自监督骨表面重建
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-19 DOI: 10.1016/j.compmedimag.2025.102690
Luohong Wu , Matthias Seibold , Nicola A. Cavalcanti , Giuseppe Loggia , Lisa Reissner , Bastian Sigrist , Jonas Hein , Lilian Calvet , Arnd Viehöfer , Philipp Fürnstahl

Background:

Bone surface reconstruction is an essential component of computer-assisted orthopedic surgery (CAOS), forming the foundation for both preoperative planning and intraoperative guidance. Compared to traditional imaging modalities such as computed tomography (CT) and magnetic resonance imaging (MRI), ultrasound, an emerging CAOS technology, provides a radiation-free, cost-effective, and portable alternative. While ultrasound offers new opportunities in CAOS, technical shortcomings continue to hinder its translation into surgery. In particular, due to the inherent limitations of ultrasound imaging, B-mode ultrasound typically captures only partial bone surfaces. The inter- and intra-operator variability in ultrasound scanning further increases the complexity of the data. Existing reconstruction methods struggle with such challenging data, leading to increased reconstruction errors and artifacts, such as holes and inflated structures. Effective techniques for accurately reconstructing open bone surfaces from real-world 3D ultrasound volumes remain lacking.

Methods:

We propose UltraBoneUDF, a self-supervised framework specifically designed for reconstructing open bone surfaces from ultrasound data. It learns unsigned distance functions (UDFs) from 3D ultrasound data. In addition, we present a novel loss function based on local tangent plane optimization that substantially improves surface reconstruction quality. UltraBoneUDF and competing models are benchmarked on three open-source datasets and further evaluated through ablation studies.

Results:

Qualitative results demonstrate the limitations of the state-of-the-art methods. Quantitatively, UltraBoneUDF achieves comparable or lower bi-directional Chamfer distance across three datasets with fewer parameters: 1.60 mm on the UltraBones100k dataset (25.5% improvement), 0.21 mm on the OpenBoneCT dataset, and 0.18 mm on the ClosedBoneCT dataset.

Conclusion:

UltraBoneUDF represents a promising solution for open bone surface reconstruction from 3D ultrasound volumes, with the potential to advance downstream applications in CAOS.
背景:骨表面重建是计算机辅助骨科手术(CAOS)的重要组成部分,是术前规划和术中指导的基础。与计算机断层扫描(CT)和磁共振成像(MRI)等传统成像方式相比,超声作为一种新兴的CAOS技术,提供了一种无辐射、成本效益高、便携的替代方案。虽然超声为CAOS提供了新的机会,但技术缺陷继续阻碍其向手术的转化。特别是,由于超声成像的固有局限性,b超通常只能捕获部分骨表面。超声扫描中操作员之间和操作员内部的可变性进一步增加了数据的复杂性。现有的重建方法难以处理这些具有挑战性的数据,导致重建误差和伪影增加,例如孔洞和膨胀结构。从真实世界的三维超声体积中准确重建开放骨表面的有效技术仍然缺乏。方法:我们提出了UltraBoneUDF,一个专门用于从超声数据重建开放骨表面的自监督框架。它从3D超声数据中学习无符号距离函数(udf)。此外,我们提出了一种新的基于局部切平面优化的损失函数,大大提高了表面重建的质量。UltraBoneUDF和竞争模型在三个开源数据集上进行基准测试,并通过消融研究进一步评估。结果:定性结果表明了最先进方法的局限性。在定量上,UltraBoneUDF在三个参数较少的数据集上实现了相当或更低的双向倒角距离:UltraBones100k数据集上的倒角距离为1.60 mm(≈25.5%),OpenBoneCT数据集上的倒角距离为0.21 mm, closebonect数据集上的倒角距离为0.18 mm。结论:UltraBoneUDF是一种很有前途的解决方案,可以从3D超声体积中重建开放骨表面,具有推进CAOS下游应用的潜力。
{"title":"UltraBoneUDF: Self-supervised bone surface reconstruction from ultrasound based on neural unsigned distance functions","authors":"Luohong Wu ,&nbsp;Matthias Seibold ,&nbsp;Nicola A. Cavalcanti ,&nbsp;Giuseppe Loggia ,&nbsp;Lisa Reissner ,&nbsp;Bastian Sigrist ,&nbsp;Jonas Hein ,&nbsp;Lilian Calvet ,&nbsp;Arnd Viehöfer ,&nbsp;Philipp Fürnstahl","doi":"10.1016/j.compmedimag.2025.102690","DOIUrl":"10.1016/j.compmedimag.2025.102690","url":null,"abstract":"<div><h3>Background:</h3><div>Bone surface reconstruction is an essential component of computer-assisted orthopedic surgery (CAOS), forming the foundation for both preoperative planning and intraoperative guidance. Compared to traditional imaging modalities such as computed tomography (CT) and magnetic resonance imaging (MRI), ultrasound, an emerging CAOS technology, provides a radiation-free, cost-effective, and portable alternative. While ultrasound offers new opportunities in CAOS, technical shortcomings continue to hinder its translation into surgery. In particular, due to the inherent limitations of ultrasound imaging, B-mode ultrasound typically captures only partial bone surfaces. The inter- and intra-operator variability in ultrasound scanning further increases the complexity of the data. Existing reconstruction methods struggle with such challenging data, leading to increased reconstruction errors and artifacts, such as holes and inflated structures. Effective techniques for accurately reconstructing open bone surfaces from real-world 3D ultrasound volumes remain lacking.</div></div><div><h3>Methods:</h3><div>We propose UltraBoneUDF, a self-supervised framework specifically designed for reconstructing open bone surfaces from ultrasound data. It learns unsigned distance functions (UDFs) from 3D ultrasound data. In addition, we present a novel loss function based on local tangent plane optimization that substantially improves surface reconstruction quality. UltraBoneUDF and competing models are benchmarked on three open-source datasets and further evaluated through ablation studies.</div></div><div><h3>Results:</h3><div>Qualitative results demonstrate the limitations of the state-of-the-art methods. Quantitatively, UltraBoneUDF achieves comparable or lower bi-directional Chamfer distance across three datasets with fewer parameters: 1.60 mm on the UltraBones100k dataset (<span><math><mrow><mo>≈</mo><mn>25</mn><mo>.</mo><mn>5</mn><mtext>%</mtext></mrow></math></span> improvement), 0.21 mm on the OpenBoneCT dataset, and 0.18 mm on the ClosedBoneCT dataset.</div></div><div><h3>Conclusion:</h3><div>UltraBoneUDF represents a promising solution for open bone surface reconstruction from 3D ultrasound volumes, with the potential to advance downstream applications in CAOS.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102690"},"PeriodicalIF":4.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145840092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Entropy-guided partial annotation for cross-domain rib segmentation 跨域肋骨分割的熵引导部分标注。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-19 DOI: 10.1016/j.compmedimag.2025.102689
Yuheng Yang , Kun You , Haoyang He , Yuehua Zhang , Xue Feng , Fei Lu , Luping Fang , Yunling Wang , Qing Pan
Deep learning methods have been widely used in medical imaging, including rib segmentation. Nevertheless, their dependence on large annotated datasets poses a significant challenge, as expert-annotated rib Computed Tomography (CT) scans are notably scarce. An additional complication arises from domain shift, which often limits the direct applicability of models trained on public datasets to specific clinical tasks, thus requiring further resource-intensive annotations on target domains for adaptation. Although semi-supervised methods have been developed to mitigate annotation costs, the prevailing strategies largely remain at the sample level. This results in unavoidable redundancy within each annotated sample, making the process of labeling an entire CT scan exceedingly tedious and costly. To address these issues, we propose a semi-supervised approach named the Entropy-Guided Partial Annotation (EGPA) method for rib segmentation. This method actively identifies the most informative regions in images for annotation based on entropy metrics, thereby substantially reducing the workload for experts during both model training and cross-domain adaptation. By integrating contrastive learning, active learning, and self-training strategies, EGPA not only significantly saves annotation cost and time when training from scratch but also effectively addresses the challenges of migrating from source to target domains. On the public RibSegV2 dataset (source domain) and a private chest CT rib segmentation dataset (target domain), EGPA achieved Dice scores of 89.5 and 90.7, respectively, nearly matching the performance of fully supervised models (89.9 and 91.2) with only 19 % and 18 % of the full annotation workload. This remarkable reduction in annotation effort shortens the development timeline for reliable segmentation tools and enhances their clinical feasibility. By simplifying the creation of high-quality annotated datasets, our approach facilitates the broad deployment of rib analysis tools in varied clinical settings, promoting standardized and efficient diagnostic practices.
深度学习方法已广泛应用于医学成像,包括肋骨分割。然而,它们对大型注释数据集的依赖构成了重大挑战,因为专家注释的肋骨计算机断层扫描(CT)非常稀缺。另一个复杂性来自领域转移,这通常限制了在公共数据集上训练的模型对特定临床任务的直接适用性,因此需要对目标领域进行进一步的资源密集型注释以适应。虽然已经开发了半监督方法来降低注释成本,但主流策略在很大程度上仍然停留在样本水平。这导致在每个注释样本中不可避免的冗余,使得标记整个CT扫描的过程非常繁琐和昂贵。为了解决这些问题,我们提出了一种半监督方法,称为熵引导部分注释(EGPA)方法,用于肋骨分割。该方法基于熵度量主动识别图像中信息量最大的区域进行标注,从而大大减少了专家在模型训练和跨域自适应过程中的工作量。通过集成对比学习、主动学习和自我训练策略,EGPA不仅在从头开始训练时显著节省注释成本和时间,而且还有效地解决了从源域迁移到目标域的挑战。在公开的RibSegV2数据集(源域)和私有的胸部CT肋骨分割数据集(目标域)上,EGPA分别获得了89.5和90.7分,几乎与完全监督模型(89.9和91.2)的性能相匹配,仅占全部注释工作量的19% %和18% %。这显著减少了注释工作,缩短了可靠分割工具的开发时间,并提高了它们的临床可行性。通过简化高质量注释数据集的创建,我们的方法有助于在各种临床环境中广泛部署肋骨分析工具,促进标准化和高效的诊断实践。
{"title":"Entropy-guided partial annotation for cross-domain rib segmentation","authors":"Yuheng Yang ,&nbsp;Kun You ,&nbsp;Haoyang He ,&nbsp;Yuehua Zhang ,&nbsp;Xue Feng ,&nbsp;Fei Lu ,&nbsp;Luping Fang ,&nbsp;Yunling Wang ,&nbsp;Qing Pan","doi":"10.1016/j.compmedimag.2025.102689","DOIUrl":"10.1016/j.compmedimag.2025.102689","url":null,"abstract":"<div><div>Deep learning methods have been widely used in medical imaging, including rib segmentation. Nevertheless, their dependence on large annotated datasets poses a significant challenge, as expert-annotated rib Computed Tomography (CT) scans are notably scarce. An additional complication arises from domain shift, which often limits the direct applicability of models trained on public datasets to specific clinical tasks, thus requiring further resource-intensive annotations on target domains for adaptation. Although semi-supervised methods have been developed to mitigate annotation costs, the prevailing strategies largely remain at the sample level. This results in unavoidable redundancy within each annotated sample, making the process of labeling an entire CT scan exceedingly tedious and costly. To address these issues, we propose a semi-supervised approach named the Entropy-Guided Partial Annotation (EGPA) method for rib segmentation. This method actively identifies the most informative regions in images for annotation based on entropy metrics, thereby substantially reducing the workload for experts during both model training and cross-domain adaptation. By integrating contrastive learning, active learning, and self-training strategies, EGPA not only significantly saves annotation cost and time when training from scratch but also effectively addresses the challenges of migrating from source to target domains. On the public RibSegV2 dataset (source domain) and a private chest CT rib segmentation dataset (target domain), EGPA achieved Dice scores of 89.5 and 90.7, respectively, nearly matching the performance of fully supervised models (89.9 and 91.2) with only 19 % and 18 % of the full annotation workload. This remarkable reduction in annotation effort shortens the development timeline for reliable segmentation tools and enhances their clinical feasibility. By simplifying the creation of high-quality annotated datasets, our approach facilitates the broad deployment of rib analysis tools in varied clinical settings, promoting standardized and efficient diagnostic practices.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102689"},"PeriodicalIF":4.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145835241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Alzheimer’s disease diagnosis by hyperspherical weighted adversarial learning in open set domain adaptation 开放集域自适应超球面加权对抗学习改进阿尔茨海默病诊断。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-19 DOI: 10.1016/j.compmedimag.2025.102692
Qiongmin Zhang, Siyi Yu, Yin Shi, Xiaowei Tan
Alzheimer’s Disease (AD) is a progressive neurodegenerative disorder characterized by cognitive decline and memory loss. Magnetic Resonance Imaging (MRI) plays a key role in detecting AD in Computer-aided Diagnosis (CAD) systems. However, variations in MRI scanners and imaging protocols introduce domain shifts, which significantly degrade model performance. Additionally, CAD models may misdiagnose unfamiliar neurodegenerative diseases not represented during training. In these complex and diverse clinical scenarios, employing closed set domain adaptation methods to achieve accurate diagnosis of AD presents substantial challenges. We propose a Hyperspherical Weighted Adversarial Learning-based Open Set Domain Adaptation (HWAL-OSDA) method for AD diagnosis. We introduce a voxel-based 3D feature extraction and fusion module to effectively capture and integrate MRI spatial features and employ a Multi-scale and Dual Attention Aggregation block to focus on disease-sensitive regions. To overcome the dispersion of feature distributions in high-dimensional space, hyperspherical variational auto-encoder module is incorporated to improve the learning of latent feature representations on a hypersphere. Furthermore, the spherical angular distance-based triplet loss and margin-based loss in the cross-domain alignment and separation module enhance the separability of known classes and establish a clear decision boundary between known and unknown classes. To improve the positive transfer of known samples and reduce the negative transfer of unknown samples, we design a weighted adversarial domain adaptation module that utilizes a dynamic instance-level weighting scheme, combining the Weibull distribution with entropy. Experiments on the ADNI and PPMI datasets show that HWAL-OSDA achieves an average accuracy of 94.2%, 83.68%, and 77.83% across three-way classification tasks (AD vs. CN vs. Unk, MCI vs. CN vs. Unk, and AD vs. MCI vs. Unk tasks), outperforming traditional and state-of-the-art OSDA methods. This approach offers a practical reference for CAD of AD and other neurodegenerative diseases in open clinical settings.
阿尔茨海默病(AD)是一种以认知能力下降和记忆丧失为特征的进行性神经退行性疾病。在计算机辅助诊断(CAD)系统中,磁共振成像(MRI)在检测AD方面起着关键作用。然而,MRI扫描仪和成像协议的变化引入了域移位,这大大降低了模型的性能。此外,CAD模型可能误诊训练中未出现的不熟悉的神经退行性疾病。在这些复杂多样的临床场景中,采用闭集域自适应方法实现AD的准确诊断提出了很大的挑战。提出了一种基于超球面加权对抗学习的开集域自适应(hwalosda)的AD诊断方法。引入基于体素的三维特征提取与融合模块,有效捕获和整合MRI空间特征,采用多尺度双注意力聚合块对疾病敏感区域进行聚焦。为了克服特征分布在高维空间中的分散性,引入了超球变分自编码器模块,提高了对超球上潜在特征表示的学习。此外,跨域对准与分离模块中基于球面角距离的三重态损失和基于边缘的损失增强了已知类的可分性,并在已知和未知类之间建立了明确的决策边界。为了提高已知样本的正迁移和减少未知样本的负迁移,我们设计了一个加权的对抗域自适应模块,该模块采用动态实例级加权方案,将威布尔分布与熵相结合。在ADNI和PPMI数据集上的实验表明,hwo -OSDA在三向分类任务(AD vs. CN vs. Unk, MCI vs. CN vs. Unk, AD vs. MCI vs. Unk)上的平均准确率分别为94.2%、83.68%和77.83%,优于传统和最先进的OSDA方法。该方法可为开放式临床环境下AD及其他神经退行性疾病的CAD提供实用参考。
{"title":"Improving Alzheimer’s disease diagnosis by hyperspherical weighted adversarial learning in open set domain adaptation","authors":"Qiongmin Zhang,&nbsp;Siyi Yu,&nbsp;Yin Shi,&nbsp;Xiaowei Tan","doi":"10.1016/j.compmedimag.2025.102692","DOIUrl":"10.1016/j.compmedimag.2025.102692","url":null,"abstract":"<div><div>Alzheimer’s Disease (AD) is a progressive neurodegenerative disorder characterized by cognitive decline and memory loss. Magnetic Resonance Imaging (MRI) plays a key role in detecting AD in Computer-aided Diagnosis (CAD) systems. However, variations in MRI scanners and imaging protocols introduce domain shifts, which significantly degrade model performance. Additionally, CAD models may misdiagnose unfamiliar neurodegenerative diseases not represented during training. In these complex and diverse clinical scenarios, employing closed set domain adaptation methods to achieve accurate diagnosis of AD presents substantial challenges. We propose a Hyperspherical Weighted Adversarial Learning-based Open Set Domain Adaptation (HWAL-OSDA) method for AD diagnosis. We introduce a voxel-based 3D feature extraction and fusion module to effectively capture and integrate MRI spatial features and employ a Multi-scale and Dual Attention Aggregation block to focus on disease-sensitive regions. To overcome the dispersion of feature distributions in high-dimensional space, hyperspherical variational auto-encoder module is incorporated to improve the learning of latent feature representations on a hypersphere. Furthermore, the spherical angular distance-based triplet loss and margin-based loss in the cross-domain alignment and separation module enhance the separability of known classes and establish a clear decision boundary between known and unknown classes. To improve the positive transfer of known samples and reduce the negative transfer of unknown samples, we design a weighted adversarial domain adaptation module that utilizes a dynamic instance-level weighting scheme, combining the Weibull distribution with entropy. Experiments on the ADNI and PPMI datasets show that HWAL-OSDA achieves an average accuracy of 94.2%, 83.68%, and 77.83% across three-way classification tasks (AD vs. CN vs. Unk, MCI vs. CN vs. Unk, and AD vs. MCI vs. Unk tasks), outperforming traditional and state-of-the-art OSDA methods. This approach offers a practical reference for CAD of AD and other neurodegenerative diseases in open clinical settings.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102692"},"PeriodicalIF":4.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
KWC-YOLO: An efficient YOLO architecture for lumbar spinal stenosis grading through dynamic convolution and spatially-aware gating KWC-YOLO:一种基于动态卷积和空间感知门控的腰椎管狭窄分级的高效YOLO架构
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-17 DOI: 10.1016/j.compmedimag.2025.102685
Wei Chen , Fan Wu , Yan Guo , Weiqing Zhuang , Hua Chen , Jie Feng , Jianfeng Wu
Lumbar spinal stenosis (LSS) represents a significant global health burden, and its diagnosis from Magnetic Resonance Imaging (MRI) is often subject to inter-observer variability and time-consuming interpretation. While deep learning (DL) models offer a promising solution, they are frequently constrained by the scarcity of annotated medical data, high computational demands, and challenges in representing subtle pathological features. To address these limitations, we propose KWC-YOLO, a novel and efficient object detection framework for the automated detection and classification of lumbar central canal stenosis (LCCS) severity according to the Schizas grading criteria. Our model enhances the YOLOv11n architecture through three core innovations: (1) the integration of KernelWarehouse (KWConv), a parameter-efficient dynamic convolution mechanism that improves the feature adaptability of the detection head; (2) the introduction of a FasterGATE activation unit in the backbone to enhance non-linear representation and accelerate convergence; and (3) the implementation of a lightweight Slim-Neck structure, which optimizes the trade-off between feature fusion quality and computational cost. On a clinical lumbar spine MRI dataset, KWC-YOLO demonstrates superior performance, achieving a mean Average Precision at an IoU of 0.5 (AP50) of 86.7% and an AP95 of 63.0%. This represents a substantial improvement over the YOLOv11n baseline by 9.2 and 9.3 percentage points in AP50 and AP95 respectively, while simultaneously reducing the computational load by 36.5% to 4.0 GFLOPs. Conclusively, KWC-YOLO establishes a new benchmark for automated LCCS grading. Its compelling balance of high accuracy and computational efficiency holds the potential to alleviate the interpretative burden on radiologists, enhance reporting accuracy, and streamline clinical decision-making, ultimately leading to improved patient outcomes.
腰椎管狭窄症(LSS)是一项重大的全球健康负担,其磁共振成像(MRI)诊断往往受到观察者之间的差异和耗时的解释。虽然深度学习(DL)模型提供了一个很有前途的解决方案,但它们经常受到带注释的医疗数据稀缺、高计算需求以及在表示细微病理特征方面的挑战的限制。为了解决这些局限性,我们提出了KWC-YOLO,这是一种新颖高效的目标检测框架,用于根据Schizas分级标准自动检测和分类腰椎中央管狭窄(LCCS)的严重程度。我们的模型通过三个核心创新增强了YOLOv11n架构:(1)集成了KernelWarehouse (KWConv),这是一种参数高效的动态卷积机制,提高了检测头的特征适应性;(2)在骨干网络中引入FasterGATE激活单元,增强非线性表示,加速收敛;(3)实现了轻量化的细颈结构,优化了特征融合质量和计算成本之间的权衡。在临床腰椎MRI数据集上,KWC-YOLO表现出优异的性能,在IoU为0.5 (AP50)时的平均平均精度为86.7%,AP95为63.0%。这在AP50和AP95中分别比YOLOv11n基线提高了9.2和9.3个百分点,同时将计算负载降低了36.5%,降至4.0 GFLOPs。最后,KWC-YOLO为LCCS自动分级建立了新的基准。它在高精度和计算效率之间取得了令人信服的平衡,有可能减轻放射科医生的解释负担,提高报告准确性,简化临床决策,最终改善患者的治疗效果。
{"title":"KWC-YOLO: An efficient YOLO architecture for lumbar spinal stenosis grading through dynamic convolution and spatially-aware gating","authors":"Wei Chen ,&nbsp;Fan Wu ,&nbsp;Yan Guo ,&nbsp;Weiqing Zhuang ,&nbsp;Hua Chen ,&nbsp;Jie Feng ,&nbsp;Jianfeng Wu","doi":"10.1016/j.compmedimag.2025.102685","DOIUrl":"10.1016/j.compmedimag.2025.102685","url":null,"abstract":"<div><div>Lumbar spinal stenosis (LSS) represents a significant global health burden, and its diagnosis from Magnetic Resonance Imaging (MRI) is often subject to inter-observer variability and time-consuming interpretation. While deep learning (DL) models offer a promising solution, they are frequently constrained by the scarcity of annotated medical data, high computational demands, and challenges in representing subtle pathological features. To address these limitations, we propose KWC-YOLO, a novel and efficient object detection framework for the automated detection and classification of lumbar central canal stenosis (LCCS) severity according to the Schizas grading criteria. Our model enhances the YOLOv11n architecture through three core innovations: (1) the integration of KernelWarehouse (KWConv), a parameter-efficient dynamic convolution mechanism that improves the feature adaptability of the detection head; (2) the introduction of a FasterGATE activation unit in the backbone to enhance non-linear representation and accelerate convergence; and (3) the implementation of a lightweight Slim-Neck structure, which optimizes the trade-off between feature fusion quality and computational cost. On a clinical lumbar spine MRI dataset, KWC-YOLO demonstrates superior performance, achieving a mean Average Precision at an IoU of 0.5 (<span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>50</mn></mrow></msub></mrow></math></span>) of 86.7% and an <span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>95</mn></mrow></msub></mrow></math></span> of 63.0%. This represents a substantial improvement over the YOLOv11n baseline by 9.2 and 9.3 percentage points in <span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>50</mn></mrow></msub></mrow></math></span> and <span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>95</mn></mrow></msub></mrow></math></span> respectively, while simultaneously reducing the computational load by 36.5% to 4.0 GFLOPs. Conclusively, KWC-YOLO establishes a new benchmark for automated LCCS grading. Its compelling balance of high accuracy and computational efficiency holds the potential to alleviate the interpretative burden on radiologists, enhance reporting accuracy, and streamline clinical decision-making, ultimately leading to improved patient outcomes.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102685"},"PeriodicalIF":4.9,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145840029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SAN-Wavelet cycleGAN for anatomical structure reinforcement and tissue detail preservation in abdominal CT synthesis 基于san -小波循环gan的腹部CT合成中解剖结构增强和组织细节保存。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-08 DOI: 10.1016/j.compmedimag.2025.102686
Yueyu Huang , Lu Qiang , Wenyu Xing , Tao Jiang , Yucheng He , Yuan Liu , Gaobo Zhang , Jingxian Wang , Xiaojun Song , Yifang Li
Generating a CT image from an existing but unpaired MR image has recently become a promising objective in the radiotherapy treatment planning (RTP), whereas the unsatisfactory structure and detail accuracy of the outcome is still a challenge for its clinical application. To address this issue, this work proposed an unsupervised model called structure-adaptive normalization wavelet (SAN-Wavelet) cycleGAN for unpaired MR-to-CT image synthesis. This method not only developed a module called SAN to ensure the anatomically structural integrity, but also introduced a neighborhood detail loss (ND loss) for detail consistency penalization between different modalities. Furthermore, a high-frequency discriminator and wavelet-trans skip connection were designed to keep with the high-frequency tissue detail. Experimental findings on abdominal area demonstrated the effectiveness of SAN-Wavelet cycleGAN for unpaired MR-to-CT synthesis, with mean squared error (MSE) of 66.38, root mean squared error (RMSE) of 8.07, peak signal-to-noise ratio (PSNR) of 25.944 dB, structural similarity index (SSlM) of 0.895 and mixture perceptual similarity index (MPSIM) of 0.723. Compared to other unsupervised approaches (i.e., cycleGAN, gc-, cc-, and sc-cycleGAN), the performances of SAN-cycleGAN improved by 5–15 % in terms of the metrics above. Moreover, the dosimetric distributions of the synthesized CT and real CT in planning target volume (PTV-45Gy) and organ at risk area (bowel) were statistically consistent (Mann-Whitney U test, P > 0.05) with a mean relative error being 0.09 % and 0.24 %, showcasing a promising applicability for clinical application.
从现有的未配对的MR图像生成CT图像最近成为放射治疗计划(RTP)的一个有希望的目标,然而结果的结构和细节准确性不理想仍然是其临床应用的一个挑战。为了解决这个问题,本研究提出了一种称为结构自适应归一化小波(san -小波)循环gan的无监督模型,用于非成对的mr - ct图像合成。该方法不仅开发了一个称为SAN的模块来保证解剖结构的完整性,而且引入了邻域细节损失(ND loss)来对不同模态之间的细节一致性进行惩罚。此外,设计了高频鉴别器和小波变换跳变连接,以保持高频组织细节。在腹部区域的实验结果表明,san_wavelet - cycleGAN对非配对mr - ct合成的有效性,均方误差(MSE)为66.38,均方根误差(RMSE)为8.07,峰值信噪比(PSNR)为25.944 dB,结构相似指数(SSlM)为0.895,混合感知相似指数(MPSIM)为0.723。与其他无监督方法(即cycleGAN, gc-, cc-和sc-cycleGAN)相比,就上述指标而言,SAN-cycleGAN的性能提高了5- 15% %。此外,合成CT与真实CT在计划靶体积(PTV-45Gy)和危险器官(肠)的剂量学分布具有统计学上的一致性(Mann-Whitney U检验,P > 0.05),平均相对误差分别为0.09 %和0.24 %,具有良好的临床应用前景。
{"title":"SAN-Wavelet cycleGAN for anatomical structure reinforcement and tissue detail preservation in abdominal CT synthesis","authors":"Yueyu Huang ,&nbsp;Lu Qiang ,&nbsp;Wenyu Xing ,&nbsp;Tao Jiang ,&nbsp;Yucheng He ,&nbsp;Yuan Liu ,&nbsp;Gaobo Zhang ,&nbsp;Jingxian Wang ,&nbsp;Xiaojun Song ,&nbsp;Yifang Li","doi":"10.1016/j.compmedimag.2025.102686","DOIUrl":"10.1016/j.compmedimag.2025.102686","url":null,"abstract":"<div><div>Generating a CT image from an existing but unpaired MR image has recently become a promising objective in the radiotherapy treatment planning (RTP), whereas the unsatisfactory structure and detail accuracy of the outcome is still a challenge for its clinical application. To address this issue, this work proposed an unsupervised model called structure-adaptive normalization wavelet (SAN-Wavelet) cycleGAN for unpaired MR-to-CT image synthesis. This method not only developed a module called SAN to ensure the anatomically structural integrity, but also introduced a neighborhood detail loss (ND loss) for detail consistency penalization between different modalities. Furthermore, a high-frequency discriminator and wavelet-trans skip connection were designed to keep with the high-frequency tissue detail. Experimental findings on abdominal area demonstrated the effectiveness of SAN-Wavelet cycleGAN for unpaired MR-to-CT synthesis, with mean squared error (MSE) of 66.38, root mean squared error (RMSE) of 8.07, peak signal-to-noise ratio (PSNR) of 25.944 dB, structural similarity index (SSlM) of 0.895 and mixture perceptual similarity index (MPSIM) of 0.723. Compared to other unsupervised approaches (i.e., cycleGAN, gc-, cc-, and sc-cycleGAN), the performances of SAN-cycleGAN improved by 5–15 % in terms of the metrics above. Moreover, the dosimetric distributions of the synthesized CT and real CT in planning target volume (PTV-45Gy) and organ at risk area (bowel) were statistically consistent (Mann-Whitney U test, P &gt; 0.05) with a mean relative error being 0.09 % and 0.24 %, showcasing a promising applicability for clinical application.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102686"},"PeriodicalIF":4.9,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145758295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel virtual patient approach for cross-patient multimodal fusion in enhanced breast cancer detection 一种新的虚拟患者方法,用于增强乳腺癌检测的跨患者多模式融合。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-08 DOI: 10.1016/j.compmedimag.2025.102687
Younes Akbari , Faseela Abdullakutty , Somaya Al-Maadeed , Rafif Al Saady , Ahmed Bouridane , Rifat Hamoudi
<div><div>Multimodal medical imaging combining conventional imaging modalities such as mammography, ultrasound, and histopathology has shown significant promise for improving breast cancer detection accuracy. However, clinical implementation faces substantial challenges due to incomplete patient-matched multimodal datasets and resource constraints. Traditional approaches require complete imaging workups from individual patients, limiting their practical applicability. This study investigates whether cross-patient multimodal fusion combining imaging modalities from different patients, can provide additional diagnostic information beyond single-modality approaches. We hypothesize that leveraging complementary information from heterogeneous patient populations enhances cancer detection performance, even when modalities originate from separate individuals. We developed a novel virtual patient framework that systematically combines imaging modalities across different patients based on quality-driven selection strategies. Two training paradigms were evaluated: Fixed scenario with 1:1:1 cross-patient combinations (<span><math><mo>∼</mo></math></span>250 virtual patients), and Combinatorial scenario with systematic companion selection (<span><math><mo>∼</mo></math></span>20,000 virtual patients). Multiple fusion architectures (concatenation, attention, and averaging) were assessed, and we designed a novel co-attention mechanism that enables sophisticated cross-modal interaction through learned attention weights. These fusion networks were evaluated using histopathology (BCSS), mammography, and ultrasound (BUSI) datasets. External validation using the ICIAR2018 BACH Challenge dataset as an alternative histopathology source demonstrated the generalizability of our approach, achieving promising accuracy despite differences in staining protocols and acquisition procedures across institutions. All models were evaluated on consistent fixed test sets to ensure fair comparison. This dataset is well-suited for multiple breast cancer analysis tasks, including detection, segmentation, and Explainable Artificial Intelligence (XAI) applications. Cross-patient multimodal fusion demonstrated significant improvements over single-modality approaches. The best single modality achieved 75.36% accuracy (mammography), while the optimal fusion combination (histopathology-mammography) reached 97.10% accuracy, representing a 21.74 percentage point improvement. Comprehensive quantitative validation through silhouette analysis (score: 0.894) confirms that the observed performance improvements reflect genuine feature space structure rather than visualization artifacts. Cross-patient multimodal fusion demonstrates significant potential for enhancing breast cancer detection, particularly addressing real-world scenarios where complete patient-matched multimodal data is unavailable. This approach represents a paradigm shift toward leveraging heterogeneous information sources for impro
多模式医学成像结合传统成像方式,如乳房x线摄影、超声和组织病理学,已显示出显著的希望,以提高乳腺癌检测的准确性。然而,由于不完整的患者匹配多模式数据集和资源限制,临床实施面临着重大挑战。传统的方法需要对单个患者进行完整的影像学检查,这限制了它们的实际适用性。本研究探讨了跨患者多模态融合结合不同患者的成像模式是否可以提供单模态方法之外的额外诊断信息。我们假设,利用来自异质患者群体的互补信息可以提高癌症检测性能,即使模式来自不同的个体。我们开发了一种新的虚拟患者框架,基于质量驱动的选择策略,系统地结合不同患者的成像模式。评估了两种训练范式:固定情景与1:1:1的交叉患者组合(约250名虚拟患者),以及组合情景与系统同伴选择(约20,000名虚拟患者)。我们评估了多种融合架构(连接、注意和平均),并设计了一种新的共同注意机制,通过学习到的注意权重实现复杂的跨模态交互。使用组织病理学(BCSS)、乳房x光检查和超声(BUSI)数据集评估这些融合网络。使用ICIAR2018巴赫挑战数据集作为替代组织病理学来源的外部验证证明了我们的方法的通用性,尽管各机构的染色方案和获取程序存在差异,但仍取得了很好的准确性。所有模型都在一致的固定测试集上进行评估,以确保公平比较。该数据集非常适合多种乳腺癌分析任务,包括检测、分割和可解释人工智能(XAI)应用。与单模态入路相比,跨患者多模态融合表现出显著的改善。最佳单一模式(乳房x线摄影)的准确率为75.36%,而最佳融合组合(组织病理学-乳房x线摄影)的准确率为97.10%,提高21.74个百分点。通过剪影分析(得分:0.894)进行的综合定量验证证实,观察到的性能改进反映了真实的特征空间结构,而不是可视化伪影。跨患者多模态融合显示了增强乳腺癌检测的巨大潜力,特别是在无法获得完整的患者匹配多模态数据的现实情况下。这种方法代表了一种范式转变,即利用异构信息源来提高诊断性能。
{"title":"A novel virtual patient approach for cross-patient multimodal fusion in enhanced breast cancer detection","authors":"Younes Akbari ,&nbsp;Faseela Abdullakutty ,&nbsp;Somaya Al-Maadeed ,&nbsp;Rafif Al Saady ,&nbsp;Ahmed Bouridane ,&nbsp;Rifat Hamoudi","doi":"10.1016/j.compmedimag.2025.102687","DOIUrl":"10.1016/j.compmedimag.2025.102687","url":null,"abstract":"&lt;div&gt;&lt;div&gt;Multimodal medical imaging combining conventional imaging modalities such as mammography, ultrasound, and histopathology has shown significant promise for improving breast cancer detection accuracy. However, clinical implementation faces substantial challenges due to incomplete patient-matched multimodal datasets and resource constraints. Traditional approaches require complete imaging workups from individual patients, limiting their practical applicability. This study investigates whether cross-patient multimodal fusion combining imaging modalities from different patients, can provide additional diagnostic information beyond single-modality approaches. We hypothesize that leveraging complementary information from heterogeneous patient populations enhances cancer detection performance, even when modalities originate from separate individuals. We developed a novel virtual patient framework that systematically combines imaging modalities across different patients based on quality-driven selection strategies. Two training paradigms were evaluated: Fixed scenario with 1:1:1 cross-patient combinations (&lt;span&gt;&lt;math&gt;&lt;mo&gt;∼&lt;/mo&gt;&lt;/math&gt;&lt;/span&gt;250 virtual patients), and Combinatorial scenario with systematic companion selection (&lt;span&gt;&lt;math&gt;&lt;mo&gt;∼&lt;/mo&gt;&lt;/math&gt;&lt;/span&gt;20,000 virtual patients). Multiple fusion architectures (concatenation, attention, and averaging) were assessed, and we designed a novel co-attention mechanism that enables sophisticated cross-modal interaction through learned attention weights. These fusion networks were evaluated using histopathology (BCSS), mammography, and ultrasound (BUSI) datasets. External validation using the ICIAR2018 BACH Challenge dataset as an alternative histopathology source demonstrated the generalizability of our approach, achieving promising accuracy despite differences in staining protocols and acquisition procedures across institutions. All models were evaluated on consistent fixed test sets to ensure fair comparison. This dataset is well-suited for multiple breast cancer analysis tasks, including detection, segmentation, and Explainable Artificial Intelligence (XAI) applications. Cross-patient multimodal fusion demonstrated significant improvements over single-modality approaches. The best single modality achieved 75.36% accuracy (mammography), while the optimal fusion combination (histopathology-mammography) reached 97.10% accuracy, representing a 21.74 percentage point improvement. Comprehensive quantitative validation through silhouette analysis (score: 0.894) confirms that the observed performance improvements reflect genuine feature space structure rather than visualization artifacts. Cross-patient multimodal fusion demonstrates significant potential for enhancing breast cancer detection, particularly addressing real-world scenarios where complete patient-matched multimodal data is unavailable. This approach represents a paradigm shift toward leveraging heterogeneous information sources for impro","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102687"},"PeriodicalIF":4.9,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145716425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computerized Medical Imaging and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1