Purpose: Accurate pixel-level segmentation is essential for medical image analysis, particularly in assisting diagnosis and treatment planning. However, fully supervised learning methods rely heavily on high-quality annotated data, which are often scarce due to the high cost of manual labeling, privacy concerns, and limited availability. We aim to reduce reliance on precise annotations and improve segmentation performance under weak supervision.
Approach: We propose scribble position and temporal contrast learning (SPTCL), an innovative segmentation method that combines contrastive learning with weak supervision. Our method leverages the spatial continuity in 3D medical image volumes and the anatomical similarities across different cardiac phases to construct a contrastive learning task for robust feature representation from unlabeled data. To enhance the feature extraction capabilities, we employ a pre-trained encoder, which is initially trained on the ACDC dataset using contrastive learning to capture robust feature representations. This pre-trained encoder is then transferred to a weakly supervised segmentation network with a dual-branch decoder for further fine-tuning on the task. The predictions from both branches are fused to generate refined pseudo-labels, which are iteratively used to guide network training with only coarse scribble annotations.
Results: Experiments on the ACDC dataset show that SPTCL outperforms existing models, achieving a Dice coefficient of 90.5%, with a 2.5% improvement over the baseline and a 1.7% improvement over the latest model. Furthermore, SPTCL reduces training time by .
Conclusions: SPTCL effectively addresses the challenges of limited annotation in medical image segmentation by uniting contrastive learning with weak supervision. It demonstrates strong potential for practical deployment in clinical settings where high-quality labels are difficult to obtain.
扫码关注我们
求助内容:
应助结果提醒方式:
