Stereo camera visual SLAM with hierarchical masking and motion-state classification at outdoor construction sites containing large dynamic objects

被引:14
作者
Bao, Runqiu [1 ]
Komatsu, Ren [1 ]
Miyagusuku, Renato [2 ]
Chino, Masaki [3 ]
Yamashita, Atsushi [1 ]
Asama, Hajime [1 ]
机构
[1] Univ Tokyo, Dept Precis Engn, Tokyo, Japan
[2] Utsunomiya Univ, Dept Mech & Intelligent Engn, Utsunomiya, Tochigi, Japan
[3] HAZAMA ANDO Corp, Construct Div, Tokyo, Japan
关键词
Dynamic visual SLAM; motion segmentation; hierarchical masking; object motion-state classification; ego-motion tracking;
D O I
10.1080/01691864.2020.1869586
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
At modern construction sites, utilizing GNSS (Global Navigation Satellite System) to measure the real-time location and orientation (i.e. pose) of construction machines and navigate them is very common. However, GNSS is not always available. Replacing GNSS with on-board cameras and visual simultaneous localization and mapping (visual SLAM) to navigate the machines is a cost-effective solution. Nevertheless, at construction sites, multiple construction machines will usually work together and side-by-side, causing large dynamic occlusions in the cameras' view. Standard visual SLAM cannot handle large dynamic occlusions well. In this work, we propose a motion segmentation method to efficiently extract static parts from crowded dynamic scenes to enable robust tracking of camera ego-motion. Our method utilizes semantic information combined with object-level geometric constraints to quickly detect the static parts of the scene. Then, we perform a two-step coarse-to-fine ego-motion tracking with reference to the static parts. This leads to a novel dynamic visual SLAM formation. We test our proposals through a real implementation based on ORB-SLAM2, and datasets we collected from real construction sites. The results show that when standard visual SLAM fails, our method can still retain accurate camera ego-motion tracking in real-time. Comparing to state-of-the-art dynamic visual SLAM methods, ours shows outstanding efficiency and competitive result trajectory accuracy.
引用
收藏
页码:228 / 241
页数:14
相关论文
共 28 条
[1]  
Antonio Meggiolaro M., P 2019 INT C ADV ROB, P135
[2]  
Barnes D, 2018, IEEE INT CONF ROBOT, P1894
[3]  
Bârsan IA, 2018, IEEE INT CONF ROBOT, P7510
[4]   DynaSLAM: Tracking, Mapping, and Inpainting in Dynamic Scenes [J].
Bescos, Berta ;
Facil, Jose M. ;
Civera, Javier ;
Neira, Jose .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :4076-4083
[5]   Real-Time Dense Mapping for Self-Driving Vehicles using Fisheye Cameras [J].
Cui, Zhaopeng ;
Heng, Lionel ;
Yeo, Ye Chuan ;
Geiger, Andreas ;
Pollefeys, Marc ;
Sattler, Torsten .
2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, :6087-6093
[6]   RANDOM SAMPLE CONSENSUS - A PARADIGM FOR MODEL-FITTING WITH APPLICATIONS TO IMAGE-ANALYSIS AND AUTOMATED CARTOGRAPHY [J].
FISCHLER, MA ;
BOLLES, RC .
COMMUNICATIONS OF THE ACM, 1981, 24 (06) :381-395
[7]   Vision meets robotics: The KITTI dataset [J].
Geiger, A. ;
Lenz, P. ;
Stiller, C. ;
Urtasun, R. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) :1231-1237
[8]  
Grupp M., 2017, evo: Python package for the evaluation of odometry and slam
[9]  
He KM, 2020, IEEE T PATTERN ANAL, V42, P386, DOI [10.1109/TPAMI.2018.2844175, 10.1109/ICCV.2017.322]
[10]  
Jaimez C., 2017, P IEEE INT C ROB AUT, P3992, DOI 10.1109/ICRA.2017.7989459