Robust Vision-Based Relative-Localization Approach Using an RGB-Depth Camera and LiDAR Sensor Fusion

被引:92
作者
Song, Haryong [1 ]
Choi, Wonsub [1 ]
Kim, Haedong [1 ]
机构
[1] Korea Aerosp Res Inst, Daejeon 34133, South Korea
关键词
Depth segmentation; interacting multiple model (IMM); intermittent observation; light detection and ranging (LiDAR); localization; modified track-to-track fusion; RGB-depth (RGB-D) camera; visual tracking; KALMAN FILTER; TRACKING; SYSTEM;
D O I
10.1109/TIE.2016.2521346
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper describes a robust vision-based relative-localization approach for a moving target based on an RGB-depth (RGB-D) camera and sensor measurements from two-dimensional (2-D) light detection and ranging (LiDAR). With the proposed approach, a target's three-dimensional (3-D) and 2-D position information is measured with an RGB-D camera and LiDAR sensor, respectively, to find the location of a target by incorporating visualtracking algorithms, depth information of the structured light sensor, and a low-level vision-LiDAR fusion algorithm, e.g., extrinsic calibration. To produce 2-D location measurements, both visual-and depth-tracking approaches are introduced, utilizing an adaptive color-based particle filter (ACPF) (for visual tracking) and an interacting multiple-model (IMM) estimator with intermittent observations from depth-image segmentation (for depth image tracking). The 2-D LiDAR data enhance location measurements by replacing results from both visual and depth tracking; through this procedure, multiple LiDAR location measurements for a target are generated. To deal with these multiple-location measurements, we propose a modified track-to-track fusion scheme. The proposed approach shows robust localization results, even when one of the trackers fails. The proposed approach was compared to position data from a Vicon motion-capture system as the ground truth. The results of this evaluation demonstrate the superiority and robustness of the proposed approach.
引用
收藏
页码:3725 / 3736
页数:12
相关论文
共 33 条
[1]  
[Anonymous], 2000, Multiple View Geometry in Computer Vision
[2]  
[Anonymous], VIC MX SYST
[3]  
[Anonymous], 2013, Learning OpenCV: Computer Vision in C++ with the OpenCVLibrary
[4]   Covariance intersection-based sensor fusion for sounding rocket tracking and impact area prediction [J].
Bolzani de Campos Ferreira, Julio Cesar ;
Waldmann, Jacques .
CONTROL ENGINEERING PRACTICE, 2007, 15 (04) :389-409
[5]  
Bradski G.R., 1998, Intel Technology Journal, V2, P1, DOI DOI 10.1109/ACV.1998.732882
[6]   A biprocessor-oriented vision-based target tracking system [J].
Canals, R ;
Roussel, A ;
Famechon, JL ;
Treuillet, S .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2002, 49 (02) :500-506
[7]   Kalman Filter for Robot Vision: A Survey [J].
Chen, S. Y. .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2012, 59 (11) :4409-4420
[8]  
Cho D., 2013, P AIAA GUID NAV CONT, P6175
[9]   Histograms of oriented gradients for human detection [J].
Dalal, N ;
Triggs, B .
2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, :886-893
[10]  
English C., 2005, P 8 INT S ART INT RO, P1