3D Object Detection Method Using LiDAR Information in Multiple Frames

被引:0
作者
Kim, Jung-Un [1 ]
Min, Jihong [2 ]
Kang, Hang-Bong [1 ]
机构
[1] Catholic Univ Korea, Bucheon, Gyeonggi Do, South Korea
[2] Agcy Def Dev, Daejeon, South Korea
来源
IMAGE ANALYSIS AND PROCESSING,(ICIAP 2017), PT I | 2017年 / 10484卷
关键词
Object detection; Deep learning; Optical flow; Sensor fusion;
D O I
10.1007/978-3-319-68560-1_25
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For a safe autonomous navigation, it is important to understand the configuration of the environment and quickly, accurately grasp the information regarding the location, direction, and size of each constituent object. Recent studies on autonomous navigation were performed to not only detect and classify objects, but also to segment and evaluate their properties. However, in these studies, pre-processing was required, which incurred a considerable amount of computational cost. Moreover, the 3D shape model was further analyzed. In other words, more computation cost and computing power are required. In this study, we propose a new method for detecting and estimating the pose of a 3D object using LiDAR information via charge-coupled-device (CCD) in real-time environment. We classified objects into classes (e.g., car, pedestrian, and cyclist), and the 3D pose of an object is quickly estimated without requiring a separate 3D-shape model. From the multiple frames obtained using the LiDAR and CCD, we design a method to robustly reconstruct the 3D environment in real time by aligning the object information of the previously obtained frames with the current frame through an optical-flow method. Our method helps in complementing the limitations of CCD-based classifiers and correcting the defects by increasing the density of the 3D-LiDAR point cloud. We compared the results obtained using our method with the state-of-the-art results of the KITTI data set; which were in good agreement in terms of speed and accuracy. This comparison shows that the 3D pose of a box can be generated with better speed and accuracy using the reconstructed 3D-point-cloud clusters proposed in our method.
引用
收藏
页码:276 / 286
页数:11
相关论文
共 15 条
  • [1] [Anonymous], 2016, SUBCATEGORY AWARE CO
  • [2] Chen XZ, 2015, ADV NEUR IN, V28
  • [3] CORTES C, 1995, MACH LEARN, V20, P273, DOI 10.1023/A:1022627411411
  • [4] MARKOV RANDOM FIELD TEXTURE MODELS
    CROSS, GR
    JAIN, AK
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1983, 5 (01) : 25 - 39
  • [5] Dai J., 2016, ADV NEURAL INFORM PR, V29, P379, DOI [DOI 10.1016/J.JPOWSOUR.2007.02.075, DOI 10.48550/ARXIV.1605.06409, DOI 10.1109/CVPR.2017.690]
  • [6] The Pascal Visual Object Classes (VOC) Challenge
    Everingham, Mark
    Van Gool, Luc
    Williams, Christopher K. I.
    Winn, John
    Zisserman, Andrew
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) : 303 - 338
  • [7] Geiger Andreas, 2012, C COMP VIS PATT REC, P2
  • [8] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [9] DETERMINING OPTICAL-FLOW
    HORN, BKP
    SCHUNCK, BG
    [J]. ARTIFICIAL INTELLIGENCE, 1981, 17 (1-3) : 185 - 203
  • [10] ImageNet Classification with Deep Convolutional Neural Networks
    Krizhevsky, Alex
    Sutskever, Ilya
    Hinton, Geoffrey E.
    [J]. COMMUNICATIONS OF THE ACM, 2017, 60 (06) : 84 - 90