An Occlusion and Noise-Aware Stereo Framework Based on Light Field Imaging for Robust Disparity Estimation

被引:29
作者
Yang, Da [1 ,2 ,3 ]
Cui, Zhenglong [1 ,2 ,3 ]
Sheng, Hao [1 ,2 ,3 ]
Chen, Rongshan [1 ,2 ,3 ]
Cong, Ruixuan [1 ,2 ,3 ]
Wang, Shuai [1 ,2 ,3 ]
Xiong, Zhang [1 ,2 ,3 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[2] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
[3] Beihang Univ, Zhongfa Aviat Inst, Key Lab Data Sci & Intelligent Comp, Hangzhou 311115, Peoples R China
关键词
Stereo vision; light field; disparity estimation; multi-direction Ray-EPI; refocusing; DEPTH ESTIMATION; SYSTEM;
D O I
10.1109/TC.2023.3343098
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Stereo vision is widely studied for depth information extraction. However, occlusion and noise pose significant challenges to traditional methods due to failure in photo consistency. In this paper, an occlusion and noise-aware stereo framework named ONAF is proposed to get a robust depth estimation by integrating the advantages of correspondence cues and refocusing cues from light field (LF). ONAF consists of two special depth cue extractors: correspondence depth cue extractor (CCE) and refocusing depth cue extractor (RCE). CCE extracts accurate correspondence depth cues in occlusion areas based on multi-direction Ray-Epipolar Plane Images (Ray-EPIs) from LF, which are more robust than traditional multi-direction EPIs. RCE generates accurate refocusing depth cues in noise areas, benefitting from the many-to-one integration strategy and the directional perception of texture and occlusion based on multi-direction focal stacks from LF. Attention mechanism is introduced to complementarily fuse CCE and RCE to generate optimum depth maps. The experimental results prove the effectiveness of ONAF, which outperforms state-of-the-art disparity estimation methods, especially in occlusion and noise areas.
引用
收藏
页码:764 / 777
页数:14
相关论文
共 43 条
[1]   Self-organizing sensor networks for integrated target surveillance [J].
Biswas, Pratik K. ;
Phoha, Shashi .
IEEE TRANSACTIONS ON COMPUTERS, 2006, 55 (08) :1033-1047
[2]   Light Field Stereo Matching Using Bilateral Statistics of Surface Cameras [J].
Chen, Can ;
Lin, Haiting ;
Yu, Zhan ;
Kang, Sing Bing ;
Yu, Jingyi .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :1518-1525
[3]   Guardauto: A Decentralized Runtime Protection System for Autonomous Driving [J].
Cheng, Kun ;
Zhou, Yuan ;
Chen, Bihuan ;
Wang, Rui ;
Bai, Yuebin ;
Liu, Yang .
IEEE TRANSACTIONS ON COMPUTERS, 2021, 70 (10) :1569-1581
[4]   Combining Implicit-Explicit View Correlation for Light Field Semantic Segmentation [J].
Cong, Ruixuan ;
Yang, Da ;
Chen, Rongshan ;
Wang, Sizhe ;
Cui, Zhenglong ;
Sheng, Hao .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, :9172-9181
[5]   Energy-Efficient 3-D Data Collection forMulti-UAV Assisted Mobile Crowdsensing [J].
Fu, Luwei ;
Zhao, Zhiwei ;
Min, Geyong ;
Miao, Wang ;
Zhao, Liang ;
Huang, Wenjie .
IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (07) :2025-2038
[6]   A Novel Occlusion-Aware Vote Cost for Light Field Depth Estimation [J].
Han, Kang ;
Xiang, Wei ;
Wang, Eric ;
Huang, Tao .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (11) :8022-8035
[7]   Neural EPI-volume Networks for Shape from Light Field [J].
Heber, Stefan ;
Yu, Wei ;
Pock, Thomas .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2271-2279
[8]   Convolutional Networks for Shape from Light Field [J].
Heber, Stefan ;
Pock, Thomas .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3746-3754
[9]   A Dataset and Evaluation Methodology for Depth Estimation on 4D Light Fields [J].
Honauer, Katrin ;
Johannsen, Ole ;
Kondermann, Daniel ;
Goldluecke, Bastian .
COMPUTER VISION - ACCV 2016, PT III, 2017, 10113 :19-34