FastFusion: Deep stereo-LiDAR fusion for real-time high-precision dense depth sensing

被引:4
作者
Meng, Haitao [1 ]
Li, Changcai [2 ]
Zhong, Chonghao [3 ]
Gu, Jianfeng [1 ]
Chen, Gang [3 ]
Knoll, Alois [1 ]
机构
[1] Tech Univ Munich, Dept Informat, Munich, Germany
[2] Sch Comp Electron & Informat, Nanning, Guangxi, Peoples R China
[3] Sun Yat sen Univ, Sch Comp Sci & Engn, Guangzhou, Guangdong, Peoples R China
关键词
artificial intelligence; depth estimation;
D O I
10.1002/rob.22216
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Light detection and ranging (LiDAR) and stereo cameras are two generally used solutions for perceiving 3D information. The complementary properties of these two sensor modalities motivate a fusion to derive practicable depth sensing toward real-world applications. Promoted by deep neural network (DNN) techniques, recent works achieve superior performance on accuracy. However, the complex architecture and the sheer number of DNN parameters often lead to poor generalization capacity and non-real-time computing. In this paper, we present FastFusion, a three-stage stereo-LiDAR deep fusion scheme, which integrates the LiDAR priors into each step of classical stereo-matching taxonomy, gaining high-precision dense depth sensing in a real-time manner. We integrate stereo-LiDAR information by taking advantage of a compact binary neural network and utilize the proposed cross-based LiDAR trust aggregation to further fuse the sparse LiDAR measurements in the back-end of stereo matching. To align the photometrical of the input image and the depth of the estimation, we introduce a refinement network to guarantee consistency. More importantly, we present a graphic processing unit-based acceleration framework for providing a low-latency implementation of FastFusion, gaining both accuracy improvement and real-time responsiveness. In the experiments, we demonstrate the effectiveness and practicability of FastFusion, which obtains a significant speedup over state-of-the-art baselines while achieving comparable accuracy on depth sensing. The video demo for real-time depth estimation of FastFusion on the real-world driving scenario is available at .
引用
收藏
页码:1804 / 1816
页数:13
相关论文
共 35 条
[1]   CBMV: A Coalesced Bidirectional Matching Volume for Disparity Estimation [J].
Batsos, Konstantinos ;
Cai, Changjiang ;
Mordohai, Philippos .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :2060-2069
[2]   Pyramid Stereo Matching Network [J].
Chang, Jia-Ren ;
Chen, Yong-Sheng .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :5410-5418
[3]   GPU-Accelerated Real-Time Stereo Estimation With Binary Neural Network [J].
Chen, Gang ;
Meng, Haitao ;
Liang, Yucheng ;
Huang, Kai .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2020, 31 (12) :2896-2907
[4]  
Cheng X., 2020, ADV NEURAL INFORM PR, V33, P22158
[5]   Noise-Aware Unsupervised Deep Lidar-Stereo Fusion [J].
Cheng, Xuelian ;
Zhong, Yiran ;
Dai, Yuchao ;
Ji, Pan ;
Li, Hongdong .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :6332-6341
[6]  
Chi Li, 2020, ICMSSP 2020: Proceedings of the 2020 5th International Conference on Multimedia Systems and Signal Processing, P11, DOI 10.1145/3404716.3404721
[7]   Volumetric Propagation Network: Stereo-LiDAR Fusion for Long-Range Depth Estimation [J].
Choe, Jaesung ;
Joo, Kyungdon ;
Imtiaz, Tooba ;
Kweon, In So .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) :4672-4679
[8]  
Geiger A, 2012, PROC CVPR IEEE, P3354, DOI 10.1109/CVPR.2012.6248074
[9]   Unsupervised Monocular Depth Estimation with Left-Right Consistency [J].
Godard, Clement ;
Mac Aodha, Oisin ;
Brostow, Gabriel J. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6602-6611
[10]   Embedded real-time stereo estimation via Semi-Global Matching on the GPU [J].
Hernandez-Juarez, D. ;
Chacon, A. ;
Espinosa, A. ;
Vazquez, D. ;
Moure, J. C. ;
Lopez, A. M. .
INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE 2016 (ICCS 2016), 2016, 80 :143-153