RGBD Salient Object Detection via Deep Fusion

被引:334
作者
Qu, Liangqiong [1 ,2 ,3 ]
He, Shengfeng [4 ]
Zhang, Jiawei [1 ]
Tian, Jiandong [2 ]
Tang, Yandong [2 ]
Yang, Qingxiong [5 ]
机构
[1] City Univ Hong Kong, Dept Comp Sci, Hong Kong, Hong Kong, Peoples R China
[2] Chinese Acad Sci, Shenyang Inst Automat, State Key Lab Robot, Shenyang 110016, Peoples R China
[3] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[4] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510006, Guangdong, Peoples R China
[5] Univ Sci & Technol China, Hefei 230000, Peoples R China
关键词
RGBD saliency detection; convolutional neural network; Laplacian propagation; ATTENTION; NETWORK; MODEL;
D O I
10.1109/TIP.2017.2682981
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Numerous efforts have been made to design various low-level saliency cues for RGBD saliency detection, such as color and depth contrast features as well as background and color compactness priors. However, how these low-level saliency cues interact with each other and how they can be effectively incorporated to generate a master saliency map remain challenging problems. In this paper, we design a new convolutional neural network (CNN) to automatically learn the interaction mechanism for RGBD salient object detection. In contrast to existing works, in which raw image pixels are fed directly to the CNN, the proposed method takes advantage of the knowledge obtained in traditional saliency detection by adopting various flexible and interpretable saliency feature vectors as inputs. This guides the CNN to learn a combination of existing features to predict saliency more effectively, which presents a less complex problem than operating on the pixels directly. We then integrate a superpixel-based Laplacian propagation framework with the trained CNN to extract a spatially consistent saliency map by exploiting the intrinsic structure of the input image. Extensive quantitative and qualitative experimental evaluations on three data sets demonstrate that the proposed method consistently outperforms the state-of-the-art methods.
引用
收藏
页码:2274 / 2285
页数:12
相关论文
共 55 条
[1]   SLIC Superpixels Compared to State-of-the-Art Superpixel Methods [J].
Achanta, Radhakrishna ;
Shaji, Appu ;
Smith, Kevin ;
Lucchi, Aurelien ;
Fua, Pascal ;
Suesstrunk, Sabine .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (11) :2274-2281
[2]  
Achanta R, 2009, PROC CVPR IEEE, P1597, DOI 10.1109/CVPRW.2009.5206596
[3]   Image Segmentation by Probabilistic Bottom-Up Aggregation and Cue Integration [J].
Alpert, Sharon ;
Galun, Meirav ;
Brandt, Achi ;
Basri, Ronen .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (02) :315-327
[4]  
[Anonymous], 2004, 2004 C COMPUTER VISI
[5]  
[Anonymous], 2014, Computer Science
[6]  
[Anonymous], P INT C SIGN PROC SY
[7]  
[Anonymous], 2015, PROC CVPR IEEE
[8]  
[Anonymous], 2014, SALIENT OBJECT DETEC
[9]  
[Anonymous], 2016, DEEPLAB SEMANTIC IMA
[10]  
Banica D, 2015, PROC CVPR IEEE, P3517, DOI 10.1109/CVPR.2015.7298974