DVDS: A deep visual dynamic slam system

被引:4
|
作者
Xie, Tao [1 ]
Sun, Qihao [1 ]
Sun, Tao [1 ]
Zhang, Jinhang [1 ]
Dai, Kun [1 ]
Zhao, Lijun [1 ]
Wang, Ke [1 ]
Li, Ruifeng [1 ]
机构
[1] Harbin Inst Technol, State Key Lab Robot & Syst, Harbin 150001, Peoples R China
基金
中国国家自然科学基金;
关键词
simultaneous localization and mapping; Transformer; Deep learning; VERSATILE;
D O I
10.1016/j.eswa.2024.125438
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Simultaneous localization and mapping (SLAM) utilizing visual sensors represent an extensively investigated research area, holding significant potential for advancements in robotics and autonomous vehicular systems. Recently, dense SLAM systems underpinned by learning-based methodologies have showcased superior accuracy and robustness compared to conventional techniques. Nevertheless, contemporary learning-based SLAM systems exhibit notable discrepancies in pose estimation, particularly within dynamic environments. In addition, the constrained receptive field of convolutional features in these methods impedes their efficacy when confronted with homogeneous, texture-less images, rendering them vulnerable to noise perturbations. We develop a novel deep visual dynamic slam (DVDS) system that exploits solely static pixels within images to retrieve the camera poses. Specifically, we formulate a dynamic object exclusion mechanism that excises dynamic constituents within the scene before the optical flow computation, thus optimizing the precision of the estimation. In addition, we unveil an efficient dispersive transformer (DisFormer) that facilitates per-pixel features in assimilating long-range information from surrounding features, culminating in constructing more precise 4D correlation volumes. Building on the DisFormer, we suggest a Disformer-based gated recurrent unit (GRU) to generate a refined flow field coupled with a confidence map, which is subsequently employed by the dense bundle adjustment layer to iteratively rectify the residuals of inverse depths and associated camera poses. The global receptive field provided by the DisFormer promotes information integration from a wider contextual window, thus improving the robustness of our SLAM system. Comprehensive experiments underscore that our proposed DVDS system manifests superior efficacy compared with state-of-the-art works across both static and dynamic scenes.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Semantic visual SLAM in dynamic environment
    Shuhuan Wen
    Pengjiang Li
    Yongjie Zhao
    Hong Zhang
    Fuchun Sun
    Zhe Wang
    Autonomous Robots, 2021, 45 : 493 - 504
  • [32] Dynamic Visual SLAM Integrated with IMU for
    Peng, Zhongcui
    Cheng, Shaowu
    Li, Xiantong
    Li, Kui
    Cai, Ming
    You, Linlin
    2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, : 4247 - 4252
  • [33] MOR-SLAM: A New Visual SLAM System for Indoor Dynamic Environments Based on Mask Restoration
    Yao, Chengzhi
    Ding, Lei
    Lan, Yonghong
    MATHEMATICS, 2023, 11 (19)
  • [34] Dynamic-SLAM: Semantic monocular visual localization and mapping based on deep learning in dynamic environment
    Xiao, Linhui
    Wang, Jinge
    Qiu, Xiaosong
    Rong, Zheng
    Zou, Xudong
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 117 : 1 - 16
  • [35] SGD-SLAM: a visual SLAM system with a dynamic feature rejection strategy combining semantic and geometric information for dynamic environments
    Chen, Yuqian
    Lu, Tongwei
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2025, 36 (02)
  • [36] Semantic visual simultaneous localization and mapping (SLAM) using deep learning for dynamic scenes
    Zhang, Xiao Ya
    Abd Rahman, Abdul Hadi
    Qamar, Faizan
    PEERJ COMPUTER SCIENCE, 2023, 9
  • [37] A Robust Deep Learning Enhanced Monocular SLAM System for Dynamic Environments
    Li, Yaoqing
    Zhong, Sheng-Hua
    Li, Shuai
    Liu, Yan
    PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023, 2023, : 508 - 515
  • [38] Semantic visual simultaneous localization and mapping (SLAM) using deep learning for dynamic scenes
    Zhang X.Y.
    Rahman A.H.A.
    Qamar F.
    PeerJ Computer Science, 2023, 9
  • [39] A visual SLAM method assisted by IMU and deep learning in indoor dynamic blurred scenes
    Liu, Fengyu
    Cao, Yi
    Cheng, Xianghong
    Liu, Luhui
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2024, 35 (02)
  • [40] Deep Learning Techniques for Visual SLAM: A Survey
    Mokssit, Saad
    Licea, Daniel Bonilla
    Guermah, Bassma
    Ghogho, Mounir
    IEEE ACCESS, 2023, 11 : 20026 - 20050