YOLOv4-tiny-based robust RGB-D SLAM approach with point and surface feature fusion in complex indoor environments

被引:8
作者
Chang, Zhanyuan [1 ,2 ]
Wu, Honglin [1 ]
Li, Chuanjiang [1 ]
机构
[1] Shanghai Normal Univ, Coll Informat Mech & Elect Engn, Shanghai, Peoples R China
[2] Shanghai Normal Univ, Coll Informat Mech & Elect Engn, Shanghai 200234, Peoples R China
关键词
dynamic environment; point features; SLAM; surface features; YOLOv4-tiny; OBJECT;
D O I
10.1002/rob.22145
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Algorithm frameworks based on feature point matching are mature and widely used in simultaneous localization and mapping (SLAM). However, in the complex and changeable indoor environment, feature point matching-based SLAM currently has two major problems, namely, decreased accuracy of pose estimation due to the interference caused by dynamic objects to the SLAM system and tracking loss caused by the lack of feature points in weak texture scenes. To address these problems, herein, we present a robust and real-time RGB-D SLAM algorithm that is based on ORBSLAM3. For interference caused by indoor moving objects, we add the improved lightweight object detection network YOLOv4-tiny to detect dynamic regions, and the dynamic features in the dynamic area are then eliminated in the algorithm tracking stage. In the case of indoor weak texture scenes, while extracting point features the system extracts surface features at the same time. The framework fuses point and surface features to track camera pose. Experiments on the public TUM RGB-D data sets show that compared with the ORB-SLAM3 algorithm in highly dynamic scenes, the root mean square error (RMSE) of the absolute path error of the proposed algorithm improved by an average of 94.08%. Camera pose is tracked without loss over time. The algorithm takes an average of 34 ms to track each frame of the picture just with a CPU, which is suitably real-time and practical. The proposed algorithm is compared with other similar algorithms, and it exhibits excellent real-time performance and accuracy. We also used a Kinect camera to evaluate our algorithm in complex indoor environment, and also showed high robustness and real-time. To sum up, our algorithm can not only deal with the interference caused by dynamic objects to the system but also stably run in the open indoor weak texture scene.
引用
收藏
页码:521 / 534
页数:14
相关论文
共 33 条
[1]   Tracking an RGB-D Camera Using Points and Planes [J].
Ataer-Cansizoglu, Esra ;
Taguchi, Yuichi ;
Ramalingam, Srikumar ;
Garaas, Tyler .
2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2013, :51-58
[2]   DynaSLAM: Tracking, Mapping, and Inpainting in Dynamic Scenes [J].
Bescos, Berta ;
Facil, Jose M. ;
Civera, Javier ;
Neira, Jose .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :4076-4083
[3]  
Campos C, 2021, IEEE Transactions on Robotics: A publication of the IEEE Robotics and Automation Society, P37
[4]   Improving Visual Localization Accuracy in Dynamic Environments Based on Dynamic Region Removal [J].
Cheng, Jiyu ;
Zhang, Hong ;
Meng, Max Q. -H. .
IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2020, 17 (03) :1585-1596
[5]   Enhancing optical-flow-based control by learning visual appearance cues for flying robots [J].
de Croon, G. C. H. E. ;
De Wagter, C. ;
Seidl, T. .
NATURE MACHINE INTELLIGENCE, 2021, 3 (01) :33-+
[6]   Tinier-YOLO: A Real-Time Object Detection Method for Constrained Environments [J].
Fang, Wei ;
Wang, Lin ;
Ren, Peiming .
IEEE ACCESS, 2020, 8 :1935-1944
[7]  
Gang Sha, 2020, 2020 IEEE International Conference on Advances in Electrical Engineering and Computer Applications (AEECA), P298, DOI 10.1109/AEECA49918.2020.9213684
[8]   RGB-D SLAM Using Point-Plane Constraints for Indoor Environments [J].
Guo, Ruibin ;
Peng, Keju ;
Fan, Weihong ;
Zhai, Yongping ;
Liu, Yunhui .
SENSORS, 2019, 19 (12)
[9]   Dynamic Scene Semantics SLAM Based on Semantic Segmentation [J].
Han, Shuangquan ;
Xi, Zhihong .
IEEE ACCESS, 2020, 8 :43563-43570
[10]  
Hsiao M, 2018, IEEE INT CONF ROBOT, P6521