Deep Light-field-driven Saliency Detection from a Single View

被引:0
作者
Piao, Yongri [1 ]
Rong, Zhengkun [1 ]
Zhang, Miao [2 ,3 ]
Li, Xiao [1 ]
Lu, Huchuan [1 ]
机构
[1] Dalian Univ Technol, Sch Informat & Commun Engn, Dalian, Peoples R China
[2] Dalian Univ Technol, Key Lab Ubiquitous Network & Serv Software Liaoni, Dalian, Peoples R China
[3] Dalian Univ Technol, DUT RU Int Sch Informat & Software Engn, Dalian, Peoples R China
来源
PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE | 2019年
基金
中国国家自然科学基金;
关键词
OBJECT DETECTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Previous 2D saliency detection methods extract salient cues from a single view and directly predict the expected results. Both traditional and deeplearning-based 2D methods do not consider geometric information of 3D scenes. Therefore the relationship between scene understanding and salient objects cannot be effectively established. This limits the performance of 2D saliency detection in challenging scenes. In this paper, we show for the first time that saliency detection problem can be reformulated as two sub-problems: light field synthesis from a single view and light-field-driven saliency detection. We propose a high-quality light field synthesis network to produce reliable 4D light field information. Then we propose a novel light-field-driven saliency detection network with two purposes, that is, i) richer saliency features can be produced for effective saliency detection; ii) geometric information can be considered for integration of multi-view saliency maps in a view-wise attention fashion. The whole pipeline can be trained in an end-to-end fashion. For training our network, we introduce the largest light field dataset for saliency detection, containing 1580 light fields that cover a wide variety of challenging scenes. With this new formulation, our method is able to achieve state-of-the-art performance.
引用
收藏
页码:904 / 911
页数:8
相关论文
共 33 条
[1]  
[Anonymous], SYMPOSIUM
[2]  
[Anonymous], 2015, CVPR
[3]  
[Anonymous], 2017, CVPR, DOI DOI 10.1109/CVPR.2017.563
[4]  
[Anonymous], 2015, IJCAI
[5]  
Buehler C, 2001, COMP GRAPH, P425, DOI 10.1145/383259.383309
[6]  
Deng ZJ, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P684
[7]   Structure-measure: A New Way to Evaluate Foreground Maps [J].
Fan, Deng-Ping ;
Cheng, Ming-Ming ;
Liu, Yun ;
Li, Tao ;
Borji, Ali .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :4558-4567
[8]  
Fan DP, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P698
[9]  
Gortler S. J., 1996, Computer Graphics Proceedings. SIGGRAPH '96, P43, DOI 10.1145/237170.237200
[10]  
Guo Xinqing, 2017, ARXIV COMPUTER VISIO