Weakly-Supervised Salient Object Detection via Scribble Annotations

被引:259
作者
Zhang, Jing [1 ,3 ,4 ]
Yu, Xin [1 ,3 ,5 ]
Li, Aixuan [2 ]
Song, Peipei [1 ,4 ]
Liu, Bowen [2 ]
Dai, Yuchao [2 ]
机构
[1] Australian Natl Univ, Canberra, ACT, Australia
[2] Northwestern Polytech Univ, Xian, Peoples R China
[3] ACRV, Brisbane, Qld, Australia
[4] Data61, Eveleigh, Australia
[5] Univ Technol Sydney, ReLER, Ultimo, Australia
来源
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020) | 2020年
关键词
D O I
10.1109/CVPR42600.2020.01256
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Compared with laborious pixel-wise dense labeling, it is much easier to label data by scribbles, which only costs 1 similar to 2 seconds to label one image. However, using scribble labels to learn salient object detection has not been explored. In this paper, we propose a weakly-supervised salient object detection model to learn saliency from such annotations. In doing so, we first relabel an existing large-scale salient object detection dataset with scribbles, namely S-DUTS dataset. Since object structure and detail information is not identified by scribbles, directly training with scribble labels will lead to saliency maps of poor boundary localization. To mitigate this problem, we propose an auxiliary edge detection task to localize object edges explicitly, and a gated structure-aware loss to place constraints on the scope of structure to be recovered. Moreover, we design a scribble boosting scheme to iteratively consolidate our scribble annotations, which are then employed as supervision to learn high-quality saliency maps. As existing saliency evaluation metrics neglect to measure structure alignment of the predictions, the saliency map ranking metric may not comply with human perception. We present a new metric, termed saliency structure measure, as a complementary metric to evaluate sharpness of the prediction. Extensive experiments on six benchmark datasets demonstrate that our method not only outperforms existing weakly-supervised/unsupervised methods, but also is on par with several fully-supervised state-of-the-art models(1).
引用
收藏
页码:12543 / 12552
页数:10
相关论文
共 57 条
[1]  
[Anonymous], 2004, ACM Transactions on Graphics (SIGGRAPH), August 2004
[2]   Improved estimates of ocean heat content from 1960 to 2015 [J].
Cheng, Lijing ;
Trenberth, Kevin E. ;
Fasullo, John ;
Boyer, Tim ;
Abraham, John ;
Zhu, Jiang .
SCIENCE ADVANCES, 2017, 3 (03)
[3]   SalientShape: group saliency in image collections [J].
Cheng, Ming-Ming ;
Mitra, Niloy J. ;
Huang, Xiaolei ;
Hu, Shi-Min .
VISUAL COMPUTER, 2014, 30 (04) :443-453
[4]   BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [J].
Dai, Jifeng ;
He, Kaiming ;
Sun, Jian .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1635-1643
[5]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[6]  
Deng ZJ, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P684
[7]  
Nguyen DT, 2019, ADV NEUR IN, V32
[8]   Structure-measure: A New Way to Evaluate Foreground Maps [J].
Fan, Deng-Ping ;
Cheng, Ming-Ming ;
Liu, Yun ;
Li, Tao ;
Borji, Ali .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :4558-4567
[9]  
Fan DP, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P698
[10]   Attentive Feedback Network for Boundary-Aware Salient Object Detection [J].
Feng, Mengyang ;
Lu, Huchuan ;
Ding, Errui .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1623-1632