LIFT-SLAM: A deep-learning feature-based monocular visual SLAM method

被引:80
作者
Silva Bruno, Hudson Martins [1 ]
Colombini, Esther Luna [1 ]
机构
[1] Univ Estadual Campinas, Inst Comp, Lab Robot & Cognit Syst, Campinas, SP, Brazil
关键词
Mobile robots; Visual SLAM; Deep neural networks; Learned features; OPTIMIZATION;
D O I
10.1016/j.neucom.2021.05.027
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Simultaneous Localization and Mapping (SLAM) problem addresses the possibility of a robot to localize itself in an unknown environment and simultaneously build a consistent map of this environment. Recently, cameras have been successfully used to get the environment's features to perform SLAM, which is referred to as visual SLAM (VSLAM). However, classical VSLAM algorithms can be easily induced to fail when either the motion of the robot or the environment is too challenging. Although new approaches based on Deep Neural Networks (DNNs) have achieved promising results in VSLAM, they still are unable to outperform traditional methods. To leverage the robustness of deep learning to enhance traditional VSLAM systems, we propose to combine the potential of deep learning-based feature descriptors with the traditional geometry-based VSLAM, building a new VSLAM system called LIFT-SLAM. Experiments conducted on KITTI and Euroc datasets show that deep learning can be used to improve the performance of traditional VSLAM systems, as the proposed approach was able to achieve results comparable to the state-of-the-art while being robust to sensorial noise. We enhance the proposed VSLAM pipeline by avoiding parameter tuning for specific datasets with an adaptive approach while evaluating how transfer learning can affect the quality of the features extracted. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:97 / 110
页数:14
相关论文
共 31 条
[1]  
Balntas E., 2016, Bmvc, P3, DOI DOI 10.5244/C.30.119
[2]  
Burri M., INT J ROBOT RES
[3]   Gradient descent optimization of smoothed information retrieval metrics [J].
Chapelle, Olivier ;
Wu, Mingrui .
INFORMATION RETRIEVAL, 2010, 13 (03) :216-235
[4]   MonoSLAM: Real-time single camera SLAM [J].
Davison, Andrew J. ;
Reid, Ian D. ;
Molton, Nicholas D. ;
Stasse, Olivier .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2007, 29 (06) :1052-1067
[5]  
DeTone D., ABS181203245 CORR
[6]   SuperPoint: Self-Supervised Interest Point Detection and Description [J].
DeTone, Daniel ;
Malisiewicz, Tomasz ;
Rabinovich, Andrew .
PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, :337-349
[7]   FlowNet: Learning Optical Flow with Convolutional Networks [J].
Dosovitskiy, Alexey ;
Fischer, Philipp ;
Ilg, Eddy ;
Haeusser, Philip ;
Hazirbas, Caner ;
Golkov, Vladimir ;
van der Smagt, Patrick ;
Cremers, Daniel ;
Brox, Thomas .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2758-2766
[8]  
Engel J., 2016, ARXIV160702555
[9]   Visual Odometry Part II: Matching, Robustness, Optimization, and Applications [J].
Fraundorfer, Friedrich ;
Scaramuzza, Davide .
IEEE ROBOTICS & AUTOMATION MAGAZINE, 2012, 19 (02) :78-90
[10]   Bags of Binary Words for Fast Place Recognition in Image Sequences [J].
Galvez-Lopez, Dorian ;
Tardos, Juan D. .
IEEE TRANSACTIONS ON ROBOTICS, 2012, 28 (05) :1188-1197