Due to the low contrast between camouflaged objects and backgrounds, the diversity of object edge shapes, and occlusions in complex scenes, existing deep learning-based Camouflaged Object Detection (COD) methods still face significant challenges in achieving high-precision detection. These challenges include difficulties in extracting multi-scale detail features for small object detection, modeling global context in occluded scenarios, and accurately distinguishing the boundaries between objects and backgrounds in complex edge detection tasks.To address these issues, this paper proposes MGCF-Net (Multi-level Global Context Fusion Network), a novel approach that integrates multi-scale context learning and feature fusion. The method employs an improved Pyramid Vision Transformer (PVTv2) as the backbone, coupled with a Cross-Scale Self-Attention (CSSA) module and a Multi-scale Fusion Attention (MFA) module. A Guided Alignment Feature Module (GAFM) aligns multi-scale features, while a large-kernel convolution structure (SHRF) enhances the global context capture capability. Experimental results on several COD benchmark data sets show that the proposed method improves 2.2%, 2.1% and 4.9% in structure metric, mean enhancement metric and weighted F metric respectively compared with FEDER, which is the second best overall performance, while the mean absolute error (MAE) decreases by 21.4%. It shows significant advantages in detection accuracy and generalization performance compared with several state-of-the-art methods (SOTA). Additionally, the method demonstrates excellent generalization to related tasks, such as polyp segmentation, COVID-19, lung infection detection, and defect detection.
扫码关注我们
求助内容:
应助结果提醒方式:
