首页 > 最新文献

Computerized Medical Imaging and Graphics最新文献

英文 中文
Entropy-guided partial annotation for cross-domain rib segmentation 跨域肋骨分割的熵引导部分标注。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-19 DOI: 10.1016/j.compmedimag.2025.102689
Yuheng Yang , Kun You , Haoyang He , Yuehua Zhang , Xue Feng , Fei Lu , Luping Fang , Yunling Wang , Qing Pan
Deep learning methods have been widely used in medical imaging, including rib segmentation. Nevertheless, their dependence on large annotated datasets poses a significant challenge, as expert-annotated rib Computed Tomography (CT) scans are notably scarce. An additional complication arises from domain shift, which often limits the direct applicability of models trained on public datasets to specific clinical tasks, thus requiring further resource-intensive annotations on target domains for adaptation. Although semi-supervised methods have been developed to mitigate annotation costs, the prevailing strategies largely remain at the sample level. This results in unavoidable redundancy within each annotated sample, making the process of labeling an entire CT scan exceedingly tedious and costly. To address these issues, we propose a semi-supervised approach named the Entropy-Guided Partial Annotation (EGPA) method for rib segmentation. This method actively identifies the most informative regions in images for annotation based on entropy metrics, thereby substantially reducing the workload for experts during both model training and cross-domain adaptation. By integrating contrastive learning, active learning, and self-training strategies, EGPA not only significantly saves annotation cost and time when training from scratch but also effectively addresses the challenges of migrating from source to target domains. On the public RibSegV2 dataset (source domain) and a private chest CT rib segmentation dataset (target domain), EGPA achieved Dice scores of 89.5 and 90.7, respectively, nearly matching the performance of fully supervised models (89.9 and 91.2) with only 19 % and 18 % of the full annotation workload. This remarkable reduction in annotation effort shortens the development timeline for reliable segmentation tools and enhances their clinical feasibility. By simplifying the creation of high-quality annotated datasets, our approach facilitates the broad deployment of rib analysis tools in varied clinical settings, promoting standardized and efficient diagnostic practices.
深度学习方法已广泛应用于医学成像,包括肋骨分割。然而,它们对大型注释数据集的依赖构成了重大挑战,因为专家注释的肋骨计算机断层扫描(CT)非常稀缺。另一个复杂性来自领域转移,这通常限制了在公共数据集上训练的模型对特定临床任务的直接适用性,因此需要对目标领域进行进一步的资源密集型注释以适应。虽然已经开发了半监督方法来降低注释成本,但主流策略在很大程度上仍然停留在样本水平。这导致在每个注释样本中不可避免的冗余,使得标记整个CT扫描的过程非常繁琐和昂贵。为了解决这些问题,我们提出了一种半监督方法,称为熵引导部分注释(EGPA)方法,用于肋骨分割。该方法基于熵度量主动识别图像中信息量最大的区域进行标注,从而大大减少了专家在模型训练和跨域自适应过程中的工作量。通过集成对比学习、主动学习和自我训练策略,EGPA不仅在从头开始训练时显著节省注释成本和时间,而且还有效地解决了从源域迁移到目标域的挑战。在公开的RibSegV2数据集(源域)和私有的胸部CT肋骨分割数据集(目标域)上,EGPA分别获得了89.5和90.7分,几乎与完全监督模型(89.9和91.2)的性能相匹配,仅占全部注释工作量的19% %和18% %。这显著减少了注释工作,缩短了可靠分割工具的开发时间,并提高了它们的临床可行性。通过简化高质量注释数据集的创建,我们的方法有助于在各种临床环境中广泛部署肋骨分析工具,促进标准化和高效的诊断实践。
{"title":"Entropy-guided partial annotation for cross-domain rib segmentation","authors":"Yuheng Yang ,&nbsp;Kun You ,&nbsp;Haoyang He ,&nbsp;Yuehua Zhang ,&nbsp;Xue Feng ,&nbsp;Fei Lu ,&nbsp;Luping Fang ,&nbsp;Yunling Wang ,&nbsp;Qing Pan","doi":"10.1016/j.compmedimag.2025.102689","DOIUrl":"10.1016/j.compmedimag.2025.102689","url":null,"abstract":"<div><div>Deep learning methods have been widely used in medical imaging, including rib segmentation. Nevertheless, their dependence on large annotated datasets poses a significant challenge, as expert-annotated rib Computed Tomography (CT) scans are notably scarce. An additional complication arises from domain shift, which often limits the direct applicability of models trained on public datasets to specific clinical tasks, thus requiring further resource-intensive annotations on target domains for adaptation. Although semi-supervised methods have been developed to mitigate annotation costs, the prevailing strategies largely remain at the sample level. This results in unavoidable redundancy within each annotated sample, making the process of labeling an entire CT scan exceedingly tedious and costly. To address these issues, we propose a semi-supervised approach named the Entropy-Guided Partial Annotation (EGPA) method for rib segmentation. This method actively identifies the most informative regions in images for annotation based on entropy metrics, thereby substantially reducing the workload for experts during both model training and cross-domain adaptation. By integrating contrastive learning, active learning, and self-training strategies, EGPA not only significantly saves annotation cost and time when training from scratch but also effectively addresses the challenges of migrating from source to target domains. On the public RibSegV2 dataset (source domain) and a private chest CT rib segmentation dataset (target domain), EGPA achieved Dice scores of 89.5 and 90.7, respectively, nearly matching the performance of fully supervised models (89.9 and 91.2) with only 19 % and 18 % of the full annotation workload. This remarkable reduction in annotation effort shortens the development timeline for reliable segmentation tools and enhances their clinical feasibility. By simplifying the creation of high-quality annotated datasets, our approach facilitates the broad deployment of rib analysis tools in varied clinical settings, promoting standardized and efficient diagnostic practices.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102689"},"PeriodicalIF":4.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145835241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Alzheimer’s disease diagnosis by hyperspherical weighted adversarial learning in open set domain adaptation 开放集域自适应超球面加权对抗学习改进阿尔茨海默病诊断。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-19 DOI: 10.1016/j.compmedimag.2025.102692
Qiongmin Zhang, Siyi Yu, Yin Shi, Xiaowei Tan
Alzheimer’s Disease (AD) is a progressive neurodegenerative disorder characterized by cognitive decline and memory loss. Magnetic Resonance Imaging (MRI) plays a key role in detecting AD in Computer-aided Diagnosis (CAD) systems. However, variations in MRI scanners and imaging protocols introduce domain shifts, which significantly degrade model performance. Additionally, CAD models may misdiagnose unfamiliar neurodegenerative diseases not represented during training. In these complex and diverse clinical scenarios, employing closed set domain adaptation methods to achieve accurate diagnosis of AD presents substantial challenges. We propose a Hyperspherical Weighted Adversarial Learning-based Open Set Domain Adaptation (HWAL-OSDA) method for AD diagnosis. We introduce a voxel-based 3D feature extraction and fusion module to effectively capture and integrate MRI spatial features and employ a Multi-scale and Dual Attention Aggregation block to focus on disease-sensitive regions. To overcome the dispersion of feature distributions in high-dimensional space, hyperspherical variational auto-encoder module is incorporated to improve the learning of latent feature representations on a hypersphere. Furthermore, the spherical angular distance-based triplet loss and margin-based loss in the cross-domain alignment and separation module enhance the separability of known classes and establish a clear decision boundary between known and unknown classes. To improve the positive transfer of known samples and reduce the negative transfer of unknown samples, we design a weighted adversarial domain adaptation module that utilizes a dynamic instance-level weighting scheme, combining the Weibull distribution with entropy. Experiments on the ADNI and PPMI datasets show that HWAL-OSDA achieves an average accuracy of 94.2%, 83.68%, and 77.83% across three-way classification tasks (AD vs. CN vs. Unk, MCI vs. CN vs. Unk, and AD vs. MCI vs. Unk tasks), outperforming traditional and state-of-the-art OSDA methods. This approach offers a practical reference for CAD of AD and other neurodegenerative diseases in open clinical settings.
阿尔茨海默病(AD)是一种以认知能力下降和记忆丧失为特征的进行性神经退行性疾病。在计算机辅助诊断(CAD)系统中,磁共振成像(MRI)在检测AD方面起着关键作用。然而,MRI扫描仪和成像协议的变化引入了域移位,这大大降低了模型的性能。此外,CAD模型可能误诊训练中未出现的不熟悉的神经退行性疾病。在这些复杂多样的临床场景中,采用闭集域自适应方法实现AD的准确诊断提出了很大的挑战。提出了一种基于超球面加权对抗学习的开集域自适应(hwalosda)的AD诊断方法。引入基于体素的三维特征提取与融合模块,有效捕获和整合MRI空间特征,采用多尺度双注意力聚合块对疾病敏感区域进行聚焦。为了克服特征分布在高维空间中的分散性,引入了超球变分自编码器模块,提高了对超球上潜在特征表示的学习。此外,跨域对准与分离模块中基于球面角距离的三重态损失和基于边缘的损失增强了已知类的可分性,并在已知和未知类之间建立了明确的决策边界。为了提高已知样本的正迁移和减少未知样本的负迁移,我们设计了一个加权的对抗域自适应模块,该模块采用动态实例级加权方案,将威布尔分布与熵相结合。在ADNI和PPMI数据集上的实验表明,hwo -OSDA在三向分类任务(AD vs. CN vs. Unk, MCI vs. CN vs. Unk, AD vs. MCI vs. Unk)上的平均准确率分别为94.2%、83.68%和77.83%,优于传统和最先进的OSDA方法。该方法可为开放式临床环境下AD及其他神经退行性疾病的CAD提供实用参考。
{"title":"Improving Alzheimer’s disease diagnosis by hyperspherical weighted adversarial learning in open set domain adaptation","authors":"Qiongmin Zhang,&nbsp;Siyi Yu,&nbsp;Yin Shi,&nbsp;Xiaowei Tan","doi":"10.1016/j.compmedimag.2025.102692","DOIUrl":"10.1016/j.compmedimag.2025.102692","url":null,"abstract":"<div><div>Alzheimer’s Disease (AD) is a progressive neurodegenerative disorder characterized by cognitive decline and memory loss. Magnetic Resonance Imaging (MRI) plays a key role in detecting AD in Computer-aided Diagnosis (CAD) systems. However, variations in MRI scanners and imaging protocols introduce domain shifts, which significantly degrade model performance. Additionally, CAD models may misdiagnose unfamiliar neurodegenerative diseases not represented during training. In these complex and diverse clinical scenarios, employing closed set domain adaptation methods to achieve accurate diagnosis of AD presents substantial challenges. We propose a Hyperspherical Weighted Adversarial Learning-based Open Set Domain Adaptation (HWAL-OSDA) method for AD diagnosis. We introduce a voxel-based 3D feature extraction and fusion module to effectively capture and integrate MRI spatial features and employ a Multi-scale and Dual Attention Aggregation block to focus on disease-sensitive regions. To overcome the dispersion of feature distributions in high-dimensional space, hyperspherical variational auto-encoder module is incorporated to improve the learning of latent feature representations on a hypersphere. Furthermore, the spherical angular distance-based triplet loss and margin-based loss in the cross-domain alignment and separation module enhance the separability of known classes and establish a clear decision boundary between known and unknown classes. To improve the positive transfer of known samples and reduce the negative transfer of unknown samples, we design a weighted adversarial domain adaptation module that utilizes a dynamic instance-level weighting scheme, combining the Weibull distribution with entropy. Experiments on the ADNI and PPMI datasets show that HWAL-OSDA achieves an average accuracy of 94.2%, 83.68%, and 77.83% across three-way classification tasks (AD vs. CN vs. Unk, MCI vs. CN vs. Unk, and AD vs. MCI vs. Unk tasks), outperforming traditional and state-of-the-art OSDA methods. This approach offers a practical reference for CAD of AD and other neurodegenerative diseases in open clinical settings.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102692"},"PeriodicalIF":4.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
KWC-YOLO: An efficient YOLO architecture for lumbar spinal stenosis grading through dynamic convolution and spatially-aware gating KWC-YOLO:一种基于动态卷积和空间感知门控的腰椎管狭窄分级的高效YOLO架构
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-17 DOI: 10.1016/j.compmedimag.2025.102685
Wei Chen , Fan Wu , Yan Guo , Weiqing Zhuang , Hua Chen , Jie Feng , Jianfeng Wu
Lumbar spinal stenosis (LSS) represents a significant global health burden, and its diagnosis from Magnetic Resonance Imaging (MRI) is often subject to inter-observer variability and time-consuming interpretation. While deep learning (DL) models offer a promising solution, they are frequently constrained by the scarcity of annotated medical data, high computational demands, and challenges in representing subtle pathological features. To address these limitations, we propose KWC-YOLO, a novel and efficient object detection framework for the automated detection and classification of lumbar central canal stenosis (LCCS) severity according to the Schizas grading criteria. Our model enhances the YOLOv11n architecture through three core innovations: (1) the integration of KernelWarehouse (KWConv), a parameter-efficient dynamic convolution mechanism that improves the feature adaptability of the detection head; (2) the introduction of a FasterGATE activation unit in the backbone to enhance non-linear representation and accelerate convergence; and (3) the implementation of a lightweight Slim-Neck structure, which optimizes the trade-off between feature fusion quality and computational cost. On a clinical lumbar spine MRI dataset, KWC-YOLO demonstrates superior performance, achieving a mean Average Precision at an IoU of 0.5 (AP50) of 86.7% and an AP95 of 63.0%. This represents a substantial improvement over the YOLOv11n baseline by 9.2 and 9.3 percentage points in AP50 and AP95 respectively, while simultaneously reducing the computational load by 36.5% to 4.0 GFLOPs. Conclusively, KWC-YOLO establishes a new benchmark for automated LCCS grading. Its compelling balance of high accuracy and computational efficiency holds the potential to alleviate the interpretative burden on radiologists, enhance reporting accuracy, and streamline clinical decision-making, ultimately leading to improved patient outcomes.
腰椎管狭窄症(LSS)是一项重大的全球健康负担,其磁共振成像(MRI)诊断往往受到观察者之间的差异和耗时的解释。虽然深度学习(DL)模型提供了一个很有前途的解决方案,但它们经常受到带注释的医疗数据稀缺、高计算需求以及在表示细微病理特征方面的挑战的限制。为了解决这些局限性,我们提出了KWC-YOLO,这是一种新颖高效的目标检测框架,用于根据Schizas分级标准自动检测和分类腰椎中央管狭窄(LCCS)的严重程度。我们的模型通过三个核心创新增强了YOLOv11n架构:(1)集成了KernelWarehouse (KWConv),这是一种参数高效的动态卷积机制,提高了检测头的特征适应性;(2)在骨干网络中引入FasterGATE激活单元,增强非线性表示,加速收敛;(3)实现了轻量化的细颈结构,优化了特征融合质量和计算成本之间的权衡。在临床腰椎MRI数据集上,KWC-YOLO表现出优异的性能,在IoU为0.5 (AP50)时的平均平均精度为86.7%,AP95为63.0%。这在AP50和AP95中分别比YOLOv11n基线提高了9.2和9.3个百分点,同时将计算负载降低了36.5%,降至4.0 GFLOPs。最后,KWC-YOLO为LCCS自动分级建立了新的基准。它在高精度和计算效率之间取得了令人信服的平衡,有可能减轻放射科医生的解释负担,提高报告准确性,简化临床决策,最终改善患者的治疗效果。
{"title":"KWC-YOLO: An efficient YOLO architecture for lumbar spinal stenosis grading through dynamic convolution and spatially-aware gating","authors":"Wei Chen ,&nbsp;Fan Wu ,&nbsp;Yan Guo ,&nbsp;Weiqing Zhuang ,&nbsp;Hua Chen ,&nbsp;Jie Feng ,&nbsp;Jianfeng Wu","doi":"10.1016/j.compmedimag.2025.102685","DOIUrl":"10.1016/j.compmedimag.2025.102685","url":null,"abstract":"<div><div>Lumbar spinal stenosis (LSS) represents a significant global health burden, and its diagnosis from Magnetic Resonance Imaging (MRI) is often subject to inter-observer variability and time-consuming interpretation. While deep learning (DL) models offer a promising solution, they are frequently constrained by the scarcity of annotated medical data, high computational demands, and challenges in representing subtle pathological features. To address these limitations, we propose KWC-YOLO, a novel and efficient object detection framework for the automated detection and classification of lumbar central canal stenosis (LCCS) severity according to the Schizas grading criteria. Our model enhances the YOLOv11n architecture through three core innovations: (1) the integration of KernelWarehouse (KWConv), a parameter-efficient dynamic convolution mechanism that improves the feature adaptability of the detection head; (2) the introduction of a FasterGATE activation unit in the backbone to enhance non-linear representation and accelerate convergence; and (3) the implementation of a lightweight Slim-Neck structure, which optimizes the trade-off between feature fusion quality and computational cost. On a clinical lumbar spine MRI dataset, KWC-YOLO demonstrates superior performance, achieving a mean Average Precision at an IoU of 0.5 (<span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>50</mn></mrow></msub></mrow></math></span>) of 86.7% and an <span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>95</mn></mrow></msub></mrow></math></span> of 63.0%. This represents a substantial improvement over the YOLOv11n baseline by 9.2 and 9.3 percentage points in <span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>50</mn></mrow></msub></mrow></math></span> and <span><math><mrow><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mn>95</mn></mrow></msub></mrow></math></span> respectively, while simultaneously reducing the computational load by 36.5% to 4.0 GFLOPs. Conclusively, KWC-YOLO establishes a new benchmark for automated LCCS grading. Its compelling balance of high accuracy and computational efficiency holds the potential to alleviate the interpretative burden on radiologists, enhance reporting accuracy, and streamline clinical decision-making, ultimately leading to improved patient outcomes.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102685"},"PeriodicalIF":4.9,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145840029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SAN-Wavelet cycleGAN for anatomical structure reinforcement and tissue detail preservation in abdominal CT synthesis 基于san -小波循环gan的腹部CT合成中解剖结构增强和组织细节保存。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-08 DOI: 10.1016/j.compmedimag.2025.102686
Yueyu Huang , Lu Qiang , Wenyu Xing , Tao Jiang , Yucheng He , Yuan Liu , Gaobo Zhang , Jingxian Wang , Xiaojun Song , Yifang Li
Generating a CT image from an existing but unpaired MR image has recently become a promising objective in the radiotherapy treatment planning (RTP), whereas the unsatisfactory structure and detail accuracy of the outcome is still a challenge for its clinical application. To address this issue, this work proposed an unsupervised model called structure-adaptive normalization wavelet (SAN-Wavelet) cycleGAN for unpaired MR-to-CT image synthesis. This method not only developed a module called SAN to ensure the anatomically structural integrity, but also introduced a neighborhood detail loss (ND loss) for detail consistency penalization between different modalities. Furthermore, a high-frequency discriminator and wavelet-trans skip connection were designed to keep with the high-frequency tissue detail. Experimental findings on abdominal area demonstrated the effectiveness of SAN-Wavelet cycleGAN for unpaired MR-to-CT synthesis, with mean squared error (MSE) of 66.38, root mean squared error (RMSE) of 8.07, peak signal-to-noise ratio (PSNR) of 25.944 dB, structural similarity index (SSlM) of 0.895 and mixture perceptual similarity index (MPSIM) of 0.723. Compared to other unsupervised approaches (i.e., cycleGAN, gc-, cc-, and sc-cycleGAN), the performances of SAN-cycleGAN improved by 5–15 % in terms of the metrics above. Moreover, the dosimetric distributions of the synthesized CT and real CT in planning target volume (PTV-45Gy) and organ at risk area (bowel) were statistically consistent (Mann-Whitney U test, P > 0.05) with a mean relative error being 0.09 % and 0.24 %, showcasing a promising applicability for clinical application.
从现有的未配对的MR图像生成CT图像最近成为放射治疗计划(RTP)的一个有希望的目标,然而结果的结构和细节准确性不理想仍然是其临床应用的一个挑战。为了解决这个问题,本研究提出了一种称为结构自适应归一化小波(san -小波)循环gan的无监督模型,用于非成对的mr - ct图像合成。该方法不仅开发了一个称为SAN的模块来保证解剖结构的完整性,而且引入了邻域细节损失(ND loss)来对不同模态之间的细节一致性进行惩罚。此外,设计了高频鉴别器和小波变换跳变连接,以保持高频组织细节。在腹部区域的实验结果表明,san_wavelet - cycleGAN对非配对mr - ct合成的有效性,均方误差(MSE)为66.38,均方根误差(RMSE)为8.07,峰值信噪比(PSNR)为25.944 dB,结构相似指数(SSlM)为0.895,混合感知相似指数(MPSIM)为0.723。与其他无监督方法(即cycleGAN, gc-, cc-和sc-cycleGAN)相比,就上述指标而言,SAN-cycleGAN的性能提高了5- 15% %。此外,合成CT与真实CT在计划靶体积(PTV-45Gy)和危险器官(肠)的剂量学分布具有统计学上的一致性(Mann-Whitney U检验,P > 0.05),平均相对误差分别为0.09 %和0.24 %,具有良好的临床应用前景。
{"title":"SAN-Wavelet cycleGAN for anatomical structure reinforcement and tissue detail preservation in abdominal CT synthesis","authors":"Yueyu Huang ,&nbsp;Lu Qiang ,&nbsp;Wenyu Xing ,&nbsp;Tao Jiang ,&nbsp;Yucheng He ,&nbsp;Yuan Liu ,&nbsp;Gaobo Zhang ,&nbsp;Jingxian Wang ,&nbsp;Xiaojun Song ,&nbsp;Yifang Li","doi":"10.1016/j.compmedimag.2025.102686","DOIUrl":"10.1016/j.compmedimag.2025.102686","url":null,"abstract":"<div><div>Generating a CT image from an existing but unpaired MR image has recently become a promising objective in the radiotherapy treatment planning (RTP), whereas the unsatisfactory structure and detail accuracy of the outcome is still a challenge for its clinical application. To address this issue, this work proposed an unsupervised model called structure-adaptive normalization wavelet (SAN-Wavelet) cycleGAN for unpaired MR-to-CT image synthesis. This method not only developed a module called SAN to ensure the anatomically structural integrity, but also introduced a neighborhood detail loss (ND loss) for detail consistency penalization between different modalities. Furthermore, a high-frequency discriminator and wavelet-trans skip connection were designed to keep with the high-frequency tissue detail. Experimental findings on abdominal area demonstrated the effectiveness of SAN-Wavelet cycleGAN for unpaired MR-to-CT synthesis, with mean squared error (MSE) of 66.38, root mean squared error (RMSE) of 8.07, peak signal-to-noise ratio (PSNR) of 25.944 dB, structural similarity index (SSlM) of 0.895 and mixture perceptual similarity index (MPSIM) of 0.723. Compared to other unsupervised approaches (i.e., cycleGAN, gc-, cc-, and sc-cycleGAN), the performances of SAN-cycleGAN improved by 5–15 % in terms of the metrics above. Moreover, the dosimetric distributions of the synthesized CT and real CT in planning target volume (PTV-45Gy) and organ at risk area (bowel) were statistically consistent (Mann-Whitney U test, P &gt; 0.05) with a mean relative error being 0.09 % and 0.24 %, showcasing a promising applicability for clinical application.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102686"},"PeriodicalIF":4.9,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145758295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel virtual patient approach for cross-patient multimodal fusion in enhanced breast cancer detection 一种新的虚拟患者方法,用于增强乳腺癌检测的跨患者多模式融合。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-08 DOI: 10.1016/j.compmedimag.2025.102687
Younes Akbari , Faseela Abdullakutty , Somaya Al-Maadeed , Rafif Al Saady , Ahmed Bouridane , Rifat Hamoudi
<div><div>Multimodal medical imaging combining conventional imaging modalities such as mammography, ultrasound, and histopathology has shown significant promise for improving breast cancer detection accuracy. However, clinical implementation faces substantial challenges due to incomplete patient-matched multimodal datasets and resource constraints. Traditional approaches require complete imaging workups from individual patients, limiting their practical applicability. This study investigates whether cross-patient multimodal fusion combining imaging modalities from different patients, can provide additional diagnostic information beyond single-modality approaches. We hypothesize that leveraging complementary information from heterogeneous patient populations enhances cancer detection performance, even when modalities originate from separate individuals. We developed a novel virtual patient framework that systematically combines imaging modalities across different patients based on quality-driven selection strategies. Two training paradigms were evaluated: Fixed scenario with 1:1:1 cross-patient combinations (<span><math><mo>∼</mo></math></span>250 virtual patients), and Combinatorial scenario with systematic companion selection (<span><math><mo>∼</mo></math></span>20,000 virtual patients). Multiple fusion architectures (concatenation, attention, and averaging) were assessed, and we designed a novel co-attention mechanism that enables sophisticated cross-modal interaction through learned attention weights. These fusion networks were evaluated using histopathology (BCSS), mammography, and ultrasound (BUSI) datasets. External validation using the ICIAR2018 BACH Challenge dataset as an alternative histopathology source demonstrated the generalizability of our approach, achieving promising accuracy despite differences in staining protocols and acquisition procedures across institutions. All models were evaluated on consistent fixed test sets to ensure fair comparison. This dataset is well-suited for multiple breast cancer analysis tasks, including detection, segmentation, and Explainable Artificial Intelligence (XAI) applications. Cross-patient multimodal fusion demonstrated significant improvements over single-modality approaches. The best single modality achieved 75.36% accuracy (mammography), while the optimal fusion combination (histopathology-mammography) reached 97.10% accuracy, representing a 21.74 percentage point improvement. Comprehensive quantitative validation through silhouette analysis (score: 0.894) confirms that the observed performance improvements reflect genuine feature space structure rather than visualization artifacts. Cross-patient multimodal fusion demonstrates significant potential for enhancing breast cancer detection, particularly addressing real-world scenarios where complete patient-matched multimodal data is unavailable. This approach represents a paradigm shift toward leveraging heterogeneous information sources for impro
多模式医学成像结合传统成像方式,如乳房x线摄影、超声和组织病理学,已显示出显著的希望,以提高乳腺癌检测的准确性。然而,由于不完整的患者匹配多模式数据集和资源限制,临床实施面临着重大挑战。传统的方法需要对单个患者进行完整的影像学检查,这限制了它们的实际适用性。本研究探讨了跨患者多模态融合结合不同患者的成像模式是否可以提供单模态方法之外的额外诊断信息。我们假设,利用来自异质患者群体的互补信息可以提高癌症检测性能,即使模式来自不同的个体。我们开发了一种新的虚拟患者框架,基于质量驱动的选择策略,系统地结合不同患者的成像模式。评估了两种训练范式:固定情景与1:1:1的交叉患者组合(约250名虚拟患者),以及组合情景与系统同伴选择(约20,000名虚拟患者)。我们评估了多种融合架构(连接、注意和平均),并设计了一种新的共同注意机制,通过学习到的注意权重实现复杂的跨模态交互。使用组织病理学(BCSS)、乳房x光检查和超声(BUSI)数据集评估这些融合网络。使用ICIAR2018巴赫挑战数据集作为替代组织病理学来源的外部验证证明了我们的方法的通用性,尽管各机构的染色方案和获取程序存在差异,但仍取得了很好的准确性。所有模型都在一致的固定测试集上进行评估,以确保公平比较。该数据集非常适合多种乳腺癌分析任务,包括检测、分割和可解释人工智能(XAI)应用。与单模态入路相比,跨患者多模态融合表现出显著的改善。最佳单一模式(乳房x线摄影)的准确率为75.36%,而最佳融合组合(组织病理学-乳房x线摄影)的准确率为97.10%,提高21.74个百分点。通过剪影分析(得分:0.894)进行的综合定量验证证实,观察到的性能改进反映了真实的特征空间结构,而不是可视化伪影。跨患者多模态融合显示了增强乳腺癌检测的巨大潜力,特别是在无法获得完整的患者匹配多模态数据的现实情况下。这种方法代表了一种范式转变,即利用异构信息源来提高诊断性能。
{"title":"A novel virtual patient approach for cross-patient multimodal fusion in enhanced breast cancer detection","authors":"Younes Akbari ,&nbsp;Faseela Abdullakutty ,&nbsp;Somaya Al-Maadeed ,&nbsp;Rafif Al Saady ,&nbsp;Ahmed Bouridane ,&nbsp;Rifat Hamoudi","doi":"10.1016/j.compmedimag.2025.102687","DOIUrl":"10.1016/j.compmedimag.2025.102687","url":null,"abstract":"&lt;div&gt;&lt;div&gt;Multimodal medical imaging combining conventional imaging modalities such as mammography, ultrasound, and histopathology has shown significant promise for improving breast cancer detection accuracy. However, clinical implementation faces substantial challenges due to incomplete patient-matched multimodal datasets and resource constraints. Traditional approaches require complete imaging workups from individual patients, limiting their practical applicability. This study investigates whether cross-patient multimodal fusion combining imaging modalities from different patients, can provide additional diagnostic information beyond single-modality approaches. We hypothesize that leveraging complementary information from heterogeneous patient populations enhances cancer detection performance, even when modalities originate from separate individuals. We developed a novel virtual patient framework that systematically combines imaging modalities across different patients based on quality-driven selection strategies. Two training paradigms were evaluated: Fixed scenario with 1:1:1 cross-patient combinations (&lt;span&gt;&lt;math&gt;&lt;mo&gt;∼&lt;/mo&gt;&lt;/math&gt;&lt;/span&gt;250 virtual patients), and Combinatorial scenario with systematic companion selection (&lt;span&gt;&lt;math&gt;&lt;mo&gt;∼&lt;/mo&gt;&lt;/math&gt;&lt;/span&gt;20,000 virtual patients). Multiple fusion architectures (concatenation, attention, and averaging) were assessed, and we designed a novel co-attention mechanism that enables sophisticated cross-modal interaction through learned attention weights. These fusion networks were evaluated using histopathology (BCSS), mammography, and ultrasound (BUSI) datasets. External validation using the ICIAR2018 BACH Challenge dataset as an alternative histopathology source demonstrated the generalizability of our approach, achieving promising accuracy despite differences in staining protocols and acquisition procedures across institutions. All models were evaluated on consistent fixed test sets to ensure fair comparison. This dataset is well-suited for multiple breast cancer analysis tasks, including detection, segmentation, and Explainable Artificial Intelligence (XAI) applications. Cross-patient multimodal fusion demonstrated significant improvements over single-modality approaches. The best single modality achieved 75.36% accuracy (mammography), while the optimal fusion combination (histopathology-mammography) reached 97.10% accuracy, representing a 21.74 percentage point improvement. Comprehensive quantitative validation through silhouette analysis (score: 0.894) confirms that the observed performance improvements reflect genuine feature space structure rather than visualization artifacts. Cross-patient multimodal fusion demonstrates significant potential for enhancing breast cancer detection, particularly addressing real-world scenarios where complete patient-matched multimodal data is unavailable. This approach represents a paradigm shift toward leveraging heterogeneous information sources for impro","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102687"},"PeriodicalIF":4.9,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145716425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FAUDA-Net: Frequency-aware unsupervised domain adaptation network for multimodal medical image segmentation fada - net:用于多模态医学图像分割的频率感知无监督域自适应网络
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-08 DOI: 10.1016/j.compmedimag.2025.102684
Yan Liu , Yan Yang , Yongquan Jiang , Xiaole Zhao , Zhuyang Xie
Unsupervised domain adaptation (UDA) has achieved notable advances in medical image segmentation. However, the significant domain shifts between source and target data continue to pose fundamental challenges, driving an urgent demand for more robust cross-domain solutions. Current UDA methods often rely on static spatial alignment strategies that ignore the dynamic evolution of feature distributions during training, often resulting in blurred boundary delineation and loss of structural details. To address these issues, we propose a frequency-aware unsupervised domain adaptation network (FAUDA-Net), which dynamically adapts to feature distribution shifts and enhances boundary delineation in the target domain. Specifically, FAUDA-Net introduces dual-domain distribution disruption to encourage domain-invariant representations, along with frequency constraints, that leverage both phase and amplitude components to guide cross-domain adaptation. Furthermore, a frequency-aware contrastive learning mechanism aligns source and target features through channel self-attention matrices, enabling a shared pixel-wise decoder to achieve robust segmentation performance. Extensive experiments on three representative medical image datasets, MMWHS17, BraTS21, and PROMISE12, demonstrate that FAUDA-Net outperforms eight state-of-the-art methods in both overall segmentation accuracy (Dice) and boundary extraction precision (ASD), providing a reliable and effective solution for multi-modal and multi-center medical image segmentation.
无监督域自适应(UDA)在医学图像分割中取得了显著进展。然而,源数据和目标数据之间的重大领域转换继续构成根本性的挑战,推动了对更强大的跨领域解决方案的迫切需求。目前的UDA方法往往依赖于静态空间对齐策略,忽略了训练过程中特征分布的动态演变,往往导致边界描绘模糊和结构细节的丢失。为了解决这些问题,我们提出了一种频率感知的无监督域自适应网络(fada - net),该网络可以动态适应特征分布的变化,并增强目标域的边界描绘。具体来说,fada - net引入了双域分布中断,以鼓励域不变表示,以及频率约束,利用相位和幅度分量来指导跨域适应。此外,频率感知的对比学习机制通过信道自注意矩阵来对齐源和目标特征,使共享像素解码器能够实现稳健的分割性能。在MMWHS17、BraTS21和PROMISE12三个具有代表性的医学图像数据集上进行的大量实验表明,FAUDA-Net在整体分割精度(Dice)和边界提取精度(ASD)方面都优于8种最先进的方法,为多模态、多中心医学图像分割提供了可靠、有效的解决方案。
{"title":"FAUDA-Net: Frequency-aware unsupervised domain adaptation network for multimodal medical image segmentation","authors":"Yan Liu ,&nbsp;Yan Yang ,&nbsp;Yongquan Jiang ,&nbsp;Xiaole Zhao ,&nbsp;Zhuyang Xie","doi":"10.1016/j.compmedimag.2025.102684","DOIUrl":"10.1016/j.compmedimag.2025.102684","url":null,"abstract":"<div><div>Unsupervised domain adaptation (UDA) has achieved notable advances in medical image segmentation. However, the significant domain shifts between source and target data continue to pose fundamental challenges, driving an urgent demand for more robust cross-domain solutions. Current UDA methods often rely on static spatial alignment strategies that ignore the dynamic evolution of feature distributions during training, often resulting in blurred boundary delineation and loss of structural details. To address these issues, we propose a frequency-aware unsupervised domain adaptation network (FAUDA-Net), which dynamically adapts to feature distribution shifts and enhances boundary delineation in the target domain. Specifically, FAUDA-Net introduces dual-domain distribution disruption to encourage domain-invariant representations, along with frequency constraints, that leverage both phase and amplitude components to guide cross-domain adaptation. Furthermore, a frequency-aware contrastive learning mechanism aligns source and target features through channel self-attention matrices, enabling a shared pixel-wise decoder to achieve robust segmentation performance. Extensive experiments on three representative medical image datasets, MMWHS17, BraTS21, and PROMISE12, demonstrate that FAUDA-Net outperforms eight state-of-the-art methods in both overall segmentation accuracy (Dice) and boundary extraction precision (ASD), providing a reliable and effective solution for multi-modal and multi-center medical image segmentation.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102684"},"PeriodicalIF":4.9,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145737505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Frequency-spatial synergistic network for accelerated multi-contrast MRI reconstruction 加速MRI多对比重建的频率-空间协同网络
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-04 DOI: 10.1016/j.compmedimag.2025.102682
Ke Li , Dong Liang , Guoqing Chen , Qiyu Jin , Tieyong Zeng
To accelerate the Magnetic Resonance Imaging (MRI) process, multi-contrast MRI reconstruction has emerged as a mainstream approach. This technique leverages readily available auxiliary modality information to guide the high-quality reconstruction of the target modality. The core challenge of this task lies in effectively capturing the global dependencies across different contrasts and efficiently integrating complementary information from multiple contrasts. To address these challenges, we propose the Multi-Contrast Dual-Domain (MCDD) network. It leverages frequency-domain information to capture global dependencies. Specifically, the network employs the Single-Contrast Learning (SCL) module as a prior to guide the Multi-Contrast Learning (MCL) module. Both modules extract complementary global and local features through Frequency (Fre) and Spatial (Spa) extraction blocks, which are fused via an Adaptive Fusion Mechanism (AFM). Experiments on BraTS, MRBrainS, and fastMRI demonstrate that MCDD outperforms state-of-the-art methods, achieving PSNR improvements of 2.42 dB (4×), 2.11 dB (8×), 1.63 dB (16×), and 0.82 dB (32×). Qualitative results further confirm enhanced structural fidelity and reduced artifacts, validating MCDD as an effective solution for accelerated MRI reconstruction.
为了加快磁共振成像(MRI)的进程,MRI多对比重建已成为一种主流方法。该技术利用易于获得的辅助模态信息来指导目标模态的高质量重建。该任务的核心挑战在于有效地捕获不同对比之间的全局依赖关系,并有效地整合来自多个对比的互补信息。为了解决这些挑战,我们提出了多对比度双域(MCDD)网络。它利用频域信息来捕获全局依赖项。具体来说,该网络采用单对比学习(Single-Contrast Learning, SCL)模块作为前置模块来指导多对比学习(Multi-Contrast Learning, MCL)模块。两个模块都通过频率(Fre)和空间(Spa)提取块提取互补的全局和局部特征,并通过自适应融合机制(AFM)进行融合。在brat、MRBrainS和fastMRI上的实验表明,MCDD优于最先进的方法,实现了2.42 dB(4倍)、2.11 dB(8倍)、1.63 dB(16倍)和0.82 dB(32倍)的PSNR改进。定性结果进一步证实了增强的结构保真度和减少的伪影,验证了MCDD是加速MRI重建的有效解决方案。
{"title":"Frequency-spatial synergistic network for accelerated multi-contrast MRI reconstruction","authors":"Ke Li ,&nbsp;Dong Liang ,&nbsp;Guoqing Chen ,&nbsp;Qiyu Jin ,&nbsp;Tieyong Zeng","doi":"10.1016/j.compmedimag.2025.102682","DOIUrl":"10.1016/j.compmedimag.2025.102682","url":null,"abstract":"<div><div>To accelerate the Magnetic Resonance Imaging (MRI) process, multi-contrast MRI reconstruction has emerged as a mainstream approach. This technique leverages readily available auxiliary modality information to guide the high-quality reconstruction of the target modality. The core challenge of this task lies in effectively capturing the global dependencies across different contrasts and efficiently integrating complementary information from multiple contrasts. To address these challenges, we propose the Multi-Contrast Dual-Domain (MCDD) network. It leverages frequency-domain information to capture global dependencies. Specifically, the network employs the Single-Contrast Learning (SCL) module as a prior to guide the Multi-Contrast Learning (MCL) module. Both modules extract complementary global and local features through Frequency (Fre) and Spatial (Spa) extraction blocks, which are fused via an Adaptive Fusion Mechanism (AFM). Experiments on BraTS, MRBrainS, and fastMRI demonstrate that MCDD outperforms state-of-the-art methods, achieving PSNR improvements of 2.42 dB (<span><math><mrow><mn>4</mn><mo>×</mo></mrow></math></span>), 2.11 dB (<span><math><mrow><mn>8</mn><mo>×</mo></mrow></math></span>), 1.63 dB (<span><math><mrow><mn>16</mn><mo>×</mo></mrow></math></span>), and 0.82 dB (<span><math><mrow><mn>32</mn><mo>×</mo></mrow></math></span>). Qualitative results further confirm enhanced structural fidelity and reduced artifacts, validating MCDD as an effective solution for accelerated MRI reconstruction.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102682"},"PeriodicalIF":4.9,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145685379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal framework for TACE treatment response prediction in patients with hepatocellular carcinoma 肝癌患者TACE治疗反应预测的多模式框架。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-02 DOI: 10.1016/j.compmedimag.2025.102679
Pietro Maria Marvulli , Domenico Buongiorno , Gian Maria Zaccaria, Antonio Brunetti, Francescomaria Marino, Vitoantonio Bevilacqua
Transarterial chemoembolization (TACE) is a first-line treatment for intermediate-stage hepatocellular carcinoma (HCC) that can cause side effects. An accurate prediction of TACE response is important to improve clinical outcomes and avoid unnecessary toxicity. This study pursues a dual objective: to propose a standardized evaluation pipeline that enables reproducible benchmarking of state-of-the-art approaches on publicly available datasets, including both internal and external validation with public dataset, and to introduce a novel multimodal framework that integrates clinical variables, radiomic and deep features extracted from CT scans using the Vision Transformer MedViT to predict treatment response. Experiments were conducted using two publicly available datasets, the HCC-TACE-Seg, used as training and internal validation sets, and the WAW-TACE cohort, used as external validation set. The results demonstrated that the proposed method outperforms existing approaches. Independent validation on the external WAW-TACE dataset achieved promising results, confirming the robustness of the model and its potential to support treatment planning.
经动脉化疗栓塞(TACE)是可引起副作用的中期肝细胞癌(HCC)的一线治疗方法。准确预测TACE反应对于改善临床结果和避免不必要的毒性非常重要。本研究追求双重目标:提出一个标准化的评估管道,使最先进的方法能够在公开可用的数据集上进行可重复的基准测试,包括使用公共数据集进行内部和外部验证,并引入一个新的多模式框架,该框架集成了临床变量,使用Vision Transformer MedViT从CT扫描中提取的放射学和深度特征,以预测治疗反应。实验使用两个公开的数据集进行,HCC-TACE-Seg作为训练和内部验证集,WAW-TACE队列作为外部验证集。结果表明,该方法优于现有方法。在外部WAW-TACE数据集上的独立验证取得了令人鼓舞的结果,证实了该模型的鲁棒性及其支持治疗计划的潜力。
{"title":"Multimodal framework for TACE treatment response prediction in patients with hepatocellular carcinoma","authors":"Pietro Maria Marvulli ,&nbsp;Domenico Buongiorno ,&nbsp;Gian Maria Zaccaria,&nbsp;Antonio Brunetti,&nbsp;Francescomaria Marino,&nbsp;Vitoantonio Bevilacqua","doi":"10.1016/j.compmedimag.2025.102679","DOIUrl":"10.1016/j.compmedimag.2025.102679","url":null,"abstract":"<div><div>Transarterial chemoembolization (TACE) is a first-line treatment for intermediate-stage hepatocellular carcinoma (HCC) that can cause side effects. An accurate prediction of TACE response is important to improve clinical outcomes and avoid unnecessary toxicity. This study pursues a dual objective: to propose a standardized evaluation pipeline that enables reproducible benchmarking of state-of-the-art approaches on publicly available datasets, including both internal and external validation with public dataset, and to introduce a novel multimodal framework that integrates clinical variables, radiomic and deep features extracted from CT scans using the Vision Transformer MedViT to predict treatment response. Experiments were conducted using two publicly available datasets, the HCC-TACE-Seg, used as training and internal validation sets, and the WAW-TACE cohort, used as external validation set. The results demonstrated that the proposed method outperforms existing approaches. Independent validation on the external WAW-TACE dataset achieved promising results, confirming the robustness of the model and its potential to support treatment planning.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102679"},"PeriodicalIF":4.9,"publicationDate":"2025-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145679264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking pathology foundation models for predicting microsatellite instability in colorectal cancer histopathology 预测结直肠癌组织病理学微卫星不稳定性的基准病理学基础模型
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-02 DOI: 10.1016/j.compmedimag.2025.102680
Mohsin Bilal , Muhammad Aamir Gulzar , Nazish Jaffar , Abdulrahman Alabduljabbar , Youssef Altherwy , Anas Alsuhaibani , Fahdah Almarshad
The rapid evolution of pathology foundation models necessitates rigorous benchmarking for clinical tasks. We evaluated three leading foundation models, UNI, Virchow2, and CONCH, for predicting microsatellite instability status from colorectal cancer whole-slide images, an essential routine clinical test. Our comprehensive framework assessed stain, tissue, and resolution invariance using datasets from The Cancer Genome Atlas (TCGA, USA; n = 409) and Pathology Artificial Intelligence Platform (PAIP, South Korea; training n = 47, testing n = 21 and n = 78). We developed an efficient pipeline with minimal preprocessing, omitting stain normalization, color augmentation, and tumor segmentation. To improve contextual encoding, we applied a five-crop strategy per patch, averaging embeddings from the center and four peripheral crops. We compared three slide-level aggregation and four efficient adaptation strategies. CONCH, using 2-cluster aggregation and ProtoNet adaptation, achieved top balanced accuracies (0.775 and 0.778) in external validation on PAIP. Conversely, UNI, with mean-averaging aggregation and ANN adaptation, excelled in TCGA cross-validation (0.778) but not in external validation (0.764), suggesting potential overfitting. The proposed 5-Crop augmentation enhances robustness to scale in UNI and CONCH and reflects intrinsic invariance achieved by Virchow2 through large-scale pretraining. For prescreening, CONCH demonstrated specificity of 0.65 and 0.45 at sensitivities of 0.90 and 0.94, respectively, highlighting its effectiveness in identifying stable cases and minimizing number of rapid molecular tests needed. Interestingly, a fine-tuned ResNet34 adaptation achieved superior performance (0.836) in the smaller internal validation cohort, suggesting current pathology foundation models training recipes may not sufficiently generalize without task-specific fine-tuning. Interpretability analyses using CONCH’s multimodal embeddings identified plasma cells as key morphological features differentiating microsatellite instability from stability, validated by pathologists (accuracy up to 92.4 %). This study underscores the feasibility and clinical significance of adapting foundation models to enhance diagnostic efficiency and patient outcomes.
病理基础模型的快速发展需要对临床任务进行严格的基准测试。我们评估了三种主要的基础模型UNI、Virchow2和CONCH,用于预测结直肠癌全切片图像的微卫星不稳定状态,这是一项重要的常规临床测试。我们的综合框架使用来自癌症基因组图谱(TCGA,美国;n = 409)和病理学人工智能平台(PAIP,韩国;训练n = 47,测试n = 21和n = 78)的数据集评估染色、组织和分辨率不变性。我们开发了一个高效的流水线,减少了预处理,省略了染色归一化,颜色增强和肿瘤分割。为了改进上下文编码,我们应用了每个补丁的五种作物策略,平均中心和四个外围作物的嵌入。我们比较了三种幻灯片级聚合和四种有效的适应策略。CONCH采用2簇聚合和ProtoNet自适应,在PAIP的外部验证中获得了最高的平衡精度(0.775和0.778)。相反,具有平均聚合和人工神经网络适应的UNI在TCGA交叉验证中表现出色(0.778),但在外部验证中表现不佳(0.764),表明可能存在过拟合。提出的5-Crop增强方法增强了UNI和CONCH的规模鲁棒性,并反映了Virchow2通过大规模预训练实现的内在不变性。对于预筛选,CONCH的特异性分别为0.65和0.45,敏感性分别为0.90和0.94,突出了其在识别稳定病例和减少所需快速分子检测数量方面的有效性。有趣的是,经过微调的ResNet34自适应在较小的内部验证队列中获得了更好的表现(0.836),这表明如果没有针对特定任务的微调,目前的病理基础模型训练方法可能无法充分推广。使用CONCH多模态嵌入的可解释性分析确定了浆细胞是区分微卫星不稳定性和稳定性的关键形态学特征,经病理学家验证(准确率高达92.4 %)。本研究强调了调整基础模型以提高诊断效率和患者预后的可行性和临床意义。
{"title":"Benchmarking pathology foundation models for predicting microsatellite instability in colorectal cancer histopathology","authors":"Mohsin Bilal ,&nbsp;Muhammad Aamir Gulzar ,&nbsp;Nazish Jaffar ,&nbsp;Abdulrahman Alabduljabbar ,&nbsp;Youssef Altherwy ,&nbsp;Anas Alsuhaibani ,&nbsp;Fahdah Almarshad","doi":"10.1016/j.compmedimag.2025.102680","DOIUrl":"10.1016/j.compmedimag.2025.102680","url":null,"abstract":"<div><div>The rapid evolution of pathology foundation models necessitates rigorous benchmarking for clinical tasks. We evaluated three leading foundation models, UNI, Virchow2, and CONCH, for predicting microsatellite instability status from colorectal cancer whole-slide images, an essential routine clinical test. Our comprehensive framework assessed stain, tissue, and resolution invariance using datasets from The Cancer Genome Atlas (TCGA, USA; n = 409) and Pathology Artificial Intelligence Platform (PAIP, South Korea; training n = 47, testing n = 21 and n = 78). We developed an efficient pipeline with minimal preprocessing, omitting stain normalization, color augmentation, and tumor segmentation. To improve contextual encoding, we applied a five-crop strategy per patch, averaging embeddings from the center and four peripheral crops. We compared three slide-level aggregation and four efficient adaptation strategies. CONCH, using 2-cluster aggregation and ProtoNet adaptation, achieved top balanced accuracies (0.775 and 0.778) in external validation on PAIP. Conversely, UNI, with mean-averaging aggregation and ANN adaptation, excelled in TCGA cross-validation (0.778) but not in external validation (0.764), suggesting potential overfitting. The proposed 5-Crop augmentation enhances robustness to scale in UNI and CONCH and reflects intrinsic invariance achieved by Virchow2 through large-scale pretraining. For prescreening, CONCH demonstrated specificity of 0.65 and 0.45 at sensitivities of 0.90 and 0.94, respectively, highlighting its effectiveness in identifying stable cases and minimizing number of rapid molecular tests needed. Interestingly, a fine-tuned ResNet34 adaptation achieved superior performance (0.836) in the smaller internal validation cohort, suggesting current pathology foundation models training recipes may not sufficiently generalize without task-specific fine-tuning. Interpretability analyses using CONCH’s multimodal embeddings identified plasma cells as key morphological features differentiating microsatellite instability from stability, validated by pathologists (accuracy up to 92.4 %). This study underscores the feasibility and clinical significance of adapting foundation models to enhance diagnostic efficiency and patient outcomes.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102680"},"PeriodicalIF":4.9,"publicationDate":"2025-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145685281","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effective contextual feature fusion and individualized information for automated sleep staging 有效的上下文特征融合和个性化信息自动睡眠分期。
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-02 DOI: 10.1016/j.compmedimag.2025.102678
Jianguo Wei , Ronghe Chu , Wenhuan Lu , Yibing Chen
Polysomnography (PSG)-based sleep stage interpretation is crucial for diagnosing sleep disorders. Over the past decade, scholars have shown that machine learning offers a faster and more convenient alternative to manual visual interpretation of sleep stages and patterns. However, neglecting contextual features and individual case differences has hindered model application in new sleep staging cases. In this paper, we propose a sleep staging model that integrates contextual feature fusion and an individualized framework. The model incorporates weighted features from multiple epochs into the scoring process, enabling accurate scoring of 30-second epoch signals. Our individualized framework is tailored for emerging cases in real-world scenarios. It aggregates unique case information to derive individualized pseudo-labels, significantly enhancing automatic sleep staging accuracy through non-independent training. This strategy effectively addresses model degradation caused by differences between training cases and single new cases. To demonstrate our approach’s efficacy, we evaluated our automated sleep staging algorithm on the Sleep-EDF-20 and Sleep-EDF-78 datasets, achieving accuracy of 85.3% and 80.8%, respectively. Furthermore, our individualized framework achieved 79.1% accuracy on the UCD dataset. These results underscore its potential as an effective tool for sleep stage classification, supporting physicians and neurologists in diagnosing sleep disorders. The proposed framework is lightweight and suitable for integration into clinical decision support system for sleep medicine, with a clear pathway for collaboration with routine laboratory scoring processes to support practical deployment.
基于多导睡眠图(PSG)的睡眠阶段解释对于诊断睡眠障碍至关重要。在过去的十年里,学者们已经证明,机器学习提供了一种比人工视觉解释睡眠阶段和模式更快、更方便的选择。然而,忽视上下文特征和个体病例差异阻碍了模型在新的睡眠分期病例中的应用。在本文中,我们提出了一个结合上下文特征融合和个性化框架的睡眠分期模型。该模型将多个epoch的加权特征纳入评分过程,能够对30秒epoch信号进行准确评分。我们的个性化框架是为现实世界中的新情况量身定制的。它聚合独特的病例信息,得出个性化的伪标签,通过非独立训练显著提高自动睡眠分期的准确性。该策略有效地解决了由于训练案例和单个新案例之间的差异而导致的模型退化问题。为了证明我们的方法的有效性,我们在sleep - edf -20和sleep - edf -78数据集上评估了我们的自动睡眠分期算法,准确率分别达到85.3%和80.8%。此外,我们的个性化框架在UCD数据集上达到了79.1%的准确率。这些结果强调了它作为睡眠阶段分类的有效工具的潜力,支持医生和神经科医生诊断睡眠障碍。所提出的框架轻量级,适合集成到睡眠医学的临床决策支持系统中,并具有与常规实验室评分流程协作的明确途径,以支持实际部署。
{"title":"Effective contextual feature fusion and individualized information for automated sleep staging","authors":"Jianguo Wei ,&nbsp;Ronghe Chu ,&nbsp;Wenhuan Lu ,&nbsp;Yibing Chen","doi":"10.1016/j.compmedimag.2025.102678","DOIUrl":"10.1016/j.compmedimag.2025.102678","url":null,"abstract":"<div><div>Polysomnography (PSG)-based sleep stage interpretation is crucial for diagnosing sleep disorders. Over the past decade, scholars have shown that machine learning offers a faster and more convenient alternative to manual visual interpretation of sleep stages and patterns. However, neglecting contextual features and individual case differences has hindered model application in new sleep staging cases. In this paper, we propose a sleep staging model that integrates contextual feature fusion and an individualized framework. The model incorporates weighted features from multiple epochs into the scoring process, enabling accurate scoring of 30-second epoch signals. Our individualized framework is tailored for emerging cases in real-world scenarios. It aggregates unique case information to derive individualized pseudo-labels, significantly enhancing automatic sleep staging accuracy through non-independent training. This strategy effectively addresses model degradation caused by differences between training cases and single new cases. To demonstrate our approach’s efficacy, we evaluated our automated sleep staging algorithm on the Sleep-EDF-20 and Sleep-EDF-78 datasets, achieving accuracy of 85.3% and 80.8%, respectively. Furthermore, our individualized framework achieved 79.1% accuracy on the UCD dataset. These results underscore its potential as an effective tool for sleep stage classification, supporting physicians and neurologists in diagnosing sleep disorders. The proposed framework is lightweight and suitable for integration into clinical decision support system for sleep medicine, with a clear pathway for collaboration with routine laboratory scoring processes to support practical deployment.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"127 ","pages":"Article 102678"},"PeriodicalIF":4.9,"publicationDate":"2025-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145670778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computerized Medical Imaging and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1