Dynamic SLAM using video object segmentation: A low cost setup for mobile robots

被引:0
作者
Tang, Zhiheng [1 ]
Chuong Nguyen [2 ]
Muthu, Sundaram [2 ]
机构
[1] Australian Natl Univ, Coll Engn Comp & Cybernet, Canberra, ACT, Australia
[2] CSIRO, Imaging & Comp Vis Data61, Canberra, ACT, Australia
来源
2024 INTERNATIONAL CONFERENCE ON DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS, DICTA | 2024年
关键词
SLAM; Dyanmic SLAM; Video object segmentation; 3D reconstruction; Robotics; VISUAL SLAM;
D O I
10.1109/DICTA63115.2024.00066
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
SLAM approaches, whether traditional, deep learning-based, or incorporating radiance field representations, face a common challenge in handling dynamic scenes due to their assumption of a static environment. This paper introduces a dynamic SLAM approach that integrates and enhances various SLAM methodologies. The primary focus lies in utilizing video object segmentation to bolster SLAM performance in dynamic environments, while concurrently developing an affordable mobile robot system tailored for such applications. While some methods have attempted dynamic SLAM, they often rely on single-object semantic segmentation, which is effective only for known object classes with prior knowledge of their static or dynamic nature. To address this limitation, we propose incorporating video object segmentation methods into our approach, which combines segmentation and tracking. Additionally, we construct a low-cost mobile robot system for dynamic SLAM applications and curate a dataset using affordable RGB-D sensors. To demonstrate the effectiveness of our method, we conduct an experimental study using the TUM-RGBD dataset and our RoverLab dynamic SLAM dataset, showcasing an enhancement in SLAM accuracy. Our system seamlessly integrates with both classical and modern Implicit SLAM algorithms, ranging from ORB-SLAM to Point-SLAM methods.
引用
收藏
页码:403 / 410
页数:8
相关论文
共 53 条
[1]  
Athar A., 2020, arXiv
[2]   DynaSLAM: Tracking, Mapping, and Inpainting in Dynamic Scenes [J].
Bescos, Berta ;
Facil, Jose M. ;
Civera, Javier ;
Neira, Jose .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :4076-4083
[3]   CodeSLAM-Learning a Compact, Optimisable Representation for Dense Visual SLAM [J].
Bloesch, Michael ;
Czarnowski, Jan ;
Clark, Ronald ;
Leutenegger, Stefan ;
Davison, Andrew J. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :2560-2568
[4]  
Brox T, 2010, LECT NOTES COMPUT SC, V6315, P282, DOI 10.1007/978-3-642-15555-0_21
[5]   ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual-Inertial, and Multimap SLAM [J].
Campos, Carlos ;
Elvira, Richard ;
Gomez Rodriguez, Juan J. ;
Montiel, Jose M. M. ;
Tardos, Juan D. .
IEEE TRANSACTIONS ON ROBOTICS, 2021, 37 (06) :1874-1890
[6]  
Chen K., 2022, Direct lidar odometry: Fast localization with dense point clouds
[7]  
He KM, 2017, IEEE I CONF COMP VIS, P2980, DOI [10.1109/TPAMI.2018.2844175, 10.1109/ICCV.2017.322]
[8]   CLOSED-FORM SOLUTION OF ABSOLUTE ORIENTATION USING UNIT QUATERNIONS [J].
HORN, BKP .
JOURNAL OF THE OPTICAL SOCIETY OF AMERICA A-OPTICS IMAGE SCIENCE AND VISION, 1987, 4 (04) :629-642
[9]  
Hu YT, 2017, ADV NEUR IN, V30
[10]   Scribble-Supervised Video Object Segmentation [J].
Huang, Peiliang ;
Han, Junwei ;
Liu, Nian ;
Ren, Jun ;
Zhang, Dingwen .
IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2022, 9 (02) :339-353