{"title":"Stacked Cross Refinement Network for Edge-Aware Salient Object Detection","authors":"Zhe Wu, Li Su, Qingming Huang","doi":"10.1109/ICCV.2019.00736","DOIUrl":null,"url":null,"abstract":"Salient object detection is a fundamental computer vision task. The majority of existing algorithms focus on aggregating multi-level features of pre-trained convolutional neural networks. Moreover, some researchers attempt to utilize edge information for auxiliary training. However, existing edge-aware models design unidirectional frameworks which only use edge features to improve the segmentation features. Motivated by the logical interrelations between binary segmentation and edge maps, we propose a novel Stacked Cross Refinement Network (SCRN) for salient object detection in this paper. Our framework aims to simultaneously refine multi-level features of salient object detection and edge detection by stacking Cross Refinement Unit (CRU). According to the logical interrelations, the CRU designs two direction-specific integration operations, and bidirectionally passes messages between the two tasks. Incorporating the refined edge-preserving features with the typical U-Net, our model detects salient objects accurately. Extensive experiments conducted on six benchmark datasets demonstrate that our method outperforms existing state-of-the-art algorithms in both accuracy and efficiency. Besides, the attribute-based performance on the SOC dataset show that the proposed model ranks first in the majority of challenging scenes. Code can be found at https://github.com/wuzhe71/SCAN.","PeriodicalId":6728,"journal":{"name":"2019 IEEE/CVF International Conference on Computer Vision (ICCV)","volume":"9 1","pages":"7263-7272"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"277","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE/CVF International Conference on Computer Vision (ICCV)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCV.2019.00736","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 277
Abstract
Salient object detection is a fundamental computer vision task. The majority of existing algorithms focus on aggregating multi-level features of pre-trained convolutional neural networks. Moreover, some researchers attempt to utilize edge information for auxiliary training. However, existing edge-aware models design unidirectional frameworks which only use edge features to improve the segmentation features. Motivated by the logical interrelations between binary segmentation and edge maps, we propose a novel Stacked Cross Refinement Network (SCRN) for salient object detection in this paper. Our framework aims to simultaneously refine multi-level features of salient object detection and edge detection by stacking Cross Refinement Unit (CRU). According to the logical interrelations, the CRU designs two direction-specific integration operations, and bidirectionally passes messages between the two tasks. Incorporating the refined edge-preserving features with the typical U-Net, our model detects salient objects accurately. Extensive experiments conducted on six benchmark datasets demonstrate that our method outperforms existing state-of-the-art algorithms in both accuracy and efficiency. Besides, the attribute-based performance on the SOC dataset show that the proposed model ranks first in the majority of challenging scenes. Code can be found at https://github.com/wuzhe71/SCAN.
显著目标检测是一项基本的计算机视觉任务。现有的卷积神经网络算法主要集中在对预训练卷积神经网络的多层特征进行聚合。此外,一些研究者试图利用边缘信息进行辅助训练。然而,现有的边缘感知模型设计了单向框架,仅利用边缘特征来改进分割特征。基于二值分割与边缘映射之间的逻辑关系,提出了一种新的用于显著目标检测的堆叠交叉细化网络(SCRN)。我们的框架旨在通过叠加交叉细化单元(Cross Refinement Unit, CRU)来同时细化显著目标检测和边缘检测的多层次特征。CRU根据逻辑关系设计两个特定方向的集成操作,并在两个任务之间双向传递消息。该模型将改进的边缘保持特征与典型的U-Net相结合,能够准确地检测出显著目标。在六个基准数据集上进行的大量实验表明,我们的方法在准确性和效率方面都优于现有的最先进算法。此外,基于属性的SOC数据集性能表明,该模型在大多数具有挑战性的场景中排名第一。代码可以在https://github.com/wuzhe71/SCAN上找到。