DM-SLAM: A Feature-Based SLAM System for Rigid Dynamic Scenes

被引:47
作者
Cheng, Junhao [1 ,2 ]
Wang, Zhi [3 ]
Zhou, Hongyan [4 ]
Li, Li [1 ]
Yao, Jian [1 ,2 ]
机构
[1] Wuhan Univ, Sch Remote Sensing & Informat Engn, Wuhan 430070, Hubei, Peoples R China
[2] Shenzhen Jimuyida Technol Co Ltd, Shenzhen 518000, Guangdong, Peoples R China
[3] Xi An Jiao Tong Univ, Sch Software Engn, Xian 710049, Shaanxi, Peoples R China
[4] Wuhan Univ, Sch Resource & Environm Sci, Wuhan 430070, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
visual SLAM; deep learning; dynamic scenes; Mask R-CNN; optical flow; ORB-SLAM2; MONOCULAR SLAM; ODOMETRY;
D O I
10.3390/ijgi9040202
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Most Simultaneous Localization and Mapping (SLAM) methods assume that environments are static. Such a strong assumption limits the application of most visual SLAM systems. The dynamic objects will cause many wrong data associations during the SLAM process. To address this problem, a novel visual SLAM method that follows the pipeline of feature-based methods called DM-SLAM is proposed in this paper. DM-SLAM combines an instance segmentation network with optical flow information to improve the location accuracy in dynamic environments, which supports monocular, stereo, and RGB-D sensors. It consists of four modules: semantic segmentation, ego-motion estimation, dynamic point detection and a feature-based SLAM framework. The semantic segmentation module obtains pixel-wise segmentation results of potentially dynamic objects, and the ego-motion estimation module calculates the initial pose. In the third module, two different strategies are presented to detect dynamic feature points for RGB-D/stereo and monocular cases. In the first case, the feature points with depth information are reprojected to the current frame. The reprojection offset vectors are used to distinguish the dynamic points. In the other case, we utilize the epipolar constraint to accomplish this task. Furthermore, the static feature points left are fed into the fourth module. The experimental results on the public TUM and KITTI datasets demonstrate that DM-SLAM outperforms the standard visual SLAM baselines in terms of accuracy in highly dynamic environments.
引用
收藏
页数:18
相关论文
共 43 条
  • [21] Visual-inertial navigation, mapping and localization: A scalable real-time causal approach
    Jones, Eagle S.
    Soatto, Stefano
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2011, 30 (04) : 407 - 430
  • [22] Kaneko M., 2018, P IEEE C COMP VIS PA, P258
  • [23] Visual Odometry Algorithm Using an RGB-D Sensor and IMU in a Highly Dynamic Environment
    Kim, Deok-Hwa
    Han, Seung-Beom
    Kim, Jong-Hwan
    [J]. ROBOT INTELLIGENCE TECHNOLOGY ANDAPPLICATIONS 3, 2015, 345 : 11 - 26
  • [24] Klein George, 2007, P1
  • [25] Keyframe-based visual-inertial odometry using nonlinear optimization
    Leutenegger, Stefan
    Lynen, Simon
    Bosse, Michael
    Siegwart, Roland
    Furgale, Paul
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2015, 34 (03) : 314 - 334
  • [26] RGB-D SLAM in Dynamic Environments Using Static Point Weighting
    Li, Shile
    Lee, Dongheui
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2017, 2 (04): : 2263 - 2270
  • [27] Lianos K.-N., 2018, PROC EUR C COMPUT VI
  • [28] Microsoft COCO: Common Objects in Context
    Lin, Tsung-Yi
    Maire, Michael
    Belongie, Serge
    Hays, James
    Perona, Pietro
    Ramanan, Deva
    Dollar, Piotr
    Zitnick, C. Lawrence
    [J]. COMPUTER VISION - ECCV 2014, PT V, 2014, 8693 : 740 - 755
  • [29] DMS-SLAM: A General Visual SLAM System for Dynamic Scenes with Multiple Sensors
    Liu, Guihua
    Zeng, Weilin
    Feng, Bo
    Xu, Feng
    [J]. SENSORS, 2019, 19 (17)
  • [30] ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras
    Mur-Artal, Raul
    Tardos, Juan D.
    [J]. IEEE TRANSACTIONS ON ROBOTICS, 2017, 33 (05) : 1255 - 1262