DepthNet: Real-Time LiDAR Point Cloud Depth Completion for Autonomous Vehicles

被引:27
作者
Bai, Lin [1 ]
Zhao, Yiming [1 ]
Elhousni, Mahdi [1 ]
Huang, Xinming [1 ]
机构
[1] Worcester Polytech Inst, Dept Elect & Comp Engn, Worcester, MA 01609 USA
来源
IEEE ACCESS | 2020年 / 8卷
关键词
Laser radar; Three-dimensional displays; Convolution; Autonomous vehicles; Real-time systems; Neural networks; Cameras; LiDAR; point cloud; depth completion; convolutional neural network; FPGA;
D O I
10.1109/ACCESS.2020.3045681
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Autonomous vehicles rely heavily on sensors such as camera and LiDAR, which provide real-time information about their surroundings for the tasks of perception, planning and control. Typically, a LiDAR can only provide sparse point cloud owing to a limited number of scanning lines. By employing depth completion, a dense depth map can be generated by assigning each camera pixel a corresponding depth value. However, the existing depth completion convolutional neural networks are very complex that requires high-end GPUs for processing, and thus they are not applicable to real-time autonomous driving. In this article, a light-weight network is proposed for the task of LiDAR point cloud depth completion. With an astonishing 96.2% reduction in the number of parameters, it still achieves comparable performance (9.3% better in MAE but 3.9% worse in RMSE) to the state-of-the-art network. For real-time embedded platforms, depthwise separable technique is applied to both convolution and deconvolution operations and the number of parameters decreases further by a factor of 7.3, with only a small percentage increase in error performance. Moreover, a system-on-chip architecture for depth completion is developed on a PYNQ-based FPGA platform that achieves real-time processing for HDL-64E LiDAR at the speed 11.1 frame per second.
引用
收藏
页码:227825 / 227833
页数:9
相关论文
共 21 条
  • [1] Novel Visible Light Communication Assisted Perspective-Four-Line Algorithm for Indoor Localization
    Bai, Lin
    Yang, Yang
    Feng, Chunyan
    Guo, Caili
    Jia, Bowen
    [J]. 2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [2] Chodosh N., 2018, ASIAN C COMPUTER VIS, P499
  • [3] Eldesokey A., 2018, British Mach. Vision Conf. (BMVC)
  • [4] SqueezeNext: Hardware-Aware Neural Network Design
    Gholami, Amir
    Kwon, Kiseok
    Wu, Bichen
    Tai, Zizheng
    Yue, Xiangyu
    Jin, Peter
    Zhao, Sicheng
    Keutzer, Kurt
    [J]. PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, : 1719 - 1728
  • [5] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [6] In Defense of Classical Image Processing: Fast Depth Completion on the CPU
    Ku, Jason
    Harakeh, Ali
    Waslander, Steven L.
    [J]. 2018 15TH CONFERENCE ON COMPUTER AND ROBOT VISION (CRV), 2018, : 16 - 22
  • [7] Optimizing CNN-based Segmentation with Deeply Customized Convolutional and Deconvolutional Architectures on FPGA
    Liu, Shuanglong
    Fan, Hongxiang
    Niu, Xinyu
    Ng, Ho-Cheung
    Chu, Yang
    Luk, Wayne
    [J]. ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2018, 11 (03)
  • [8] Lyu Y., 2018, PLOS COMPUTATIONAL B, P1
  • [9] Ma FC, 2019, IEEE INT CONF ROBOT, P3288, DOI [10.1109/ICRA.2019.8793637, 10.1109/icra.2019.8793637]
  • [10] Optimizing Loop Operation and Dataflow in FPGA Acceleration of Deep Convolutional Neural Networks
    Ma, Yufei
    Cao, Yu
    Vrudhula, Sarma
    Seo, Jae-sun
    [J]. FPGA'17: PROCEEDINGS OF THE 2017 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS, 2017, : 45 - 54