Depth-Quality-Aware Salient Object Detection

被引:89
作者
Chen, Chenglizhao [1 ]
Wei, Jipeng [1 ]
Peng, Chong [1 ]
Qin, Hong [2 ]
机构
[1] Qingdao Univ, Coll Comp Sci & Technol, Qingdao 266071, Peoples R China
[2] SUNY Stony Brook, Dept Comp Sci, Stony Brook, NY 11794 USA
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Object detection; Feature extraction; Streaming media; Training; Deep learning; Computational modeling; Task analysis; RGB-D salient object detection; weakly supervised learning; FUSION; SEGMENTATION; NETWORK;
D O I
10.1109/TIP.2021.3052069
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The existing fusion-based RGB-D salient object detection methods usually adopt the bistream structure to strike a balance in the fusion trade-off between RGB and depth (D). While the D quality usually varies among the scenes, the state-of-the-art bistream approaches are depth-quality-unaware, resulting in substantial difficulties in achieving complementary fusion status between RGB and D and leading to poor fusion results for low-quality D. Thus, this paper attempts to integrate a novel depth-quality-aware subnet into the classic bistream structure in order to assess the depth quality prior to conducting the selective RGB-D fusion. Compared to the SOTA bistream methods, the major advantage of our method is its ability to lessen the importance of the low-quality, no-contribution, or even negative-contribution D regions during RGB-D fusion, achieving a much improved complementary status between RGB and D. Our source code and data are available online at https://github.com/qdu1995/DQSD.
引用
收藏
页码:2350 / 2363
页数:14
相关论文
共 74 条
[1]  
Achanta R, 2009, PROC CVPR IEEE, P1597, DOI 10.1109/CVPRW.2009.5206596
[2]  
Chen C., 2020, ARXIV PREPRINT ARXIV
[3]   Improved Robust Video Saliency Detection Based on Long-Term Spatial-Temporal Information [J].
Chen, Chenglizhao ;
Wang, Guotao ;
Peng, Chong ;
Zhang, Xiaowei ;
Qin, Hong .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :1090-1100
[4]   Bilevel Feature Learning for Video Saliency Detection [J].
Chen, Chenglizhao ;
Li, Shuai ;
Qin, Hong ;
Pan, Zhenkuan ;
Yang, Guowei .
IEEE TRANSACTIONS ON MULTIMEDIA, 2018, 20 (12) :3324-3336
[5]   Video Saliency Detection via Spatial-Temporal Fusion and Low-Rank Coherency Diffusion [J].
Chen, Chenglizhao ;
Li, Shuai ;
Wang, Yongguang ;
Qin, Hong ;
Hao, Aimin .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (07) :3156-3170
[6]   Robust salient motion detection in non-stationary videos via novel integrated strategies of spatio-temporal coherency clues and low-rank analysis [J].
Chen, Chenglizhao ;
Li, Shuai ;
Qin, Hong ;
Hao, Aimin .
PATTERN RECOGNITION, 2016, 52 :410-432
[7]   Real-time and robust object tracking in video via low-rank coherency analysis in feature space [J].
Chen, Chenglizhao ;
Li, Shuai ;
Qin, Hong ;
Hao, Aimin .
PATTERN RECOGNITION, 2015, 48 (09) :2885-2905
[8]   Structure-Sensitive Saliency Detection via Multilevel Rank Analysis in Intrinsic Feature Space [J].
Chen, Chenglizhao ;
Li, Shuai ;
Qin, Hong ;
Hao, Aimin .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2015, 24 (08) :2303-2316
[9]   Three-Stream Attention-Aware Network for RGB-D Salient Object Detection [J].
Chen, Hao ;
Li, Youfu .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (06) :2825-2835
[10]   Progressively Complementarity-aware Fusion Network for RGB-D Salient Object Detection [J].
Chen, Hao ;
Li, Youfu .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :3051-3060