Deep Reinforcement Learning-Based Large-Scale Robot Exploration

被引:1
作者
Cao, Yuhong [1 ]
Zhao, Rui [1 ]
Wang, Yizhuo [1 ]
Xiang, Bairan [1 ]
Sartoretti, Guillaume [1 ]
机构
[1] Natl Univ Singapore, Coll Design & Engn, Dept Mech Engn, Singapore 117482, Singapore
关键词
View Planning for SLAM; reinforcement learning; motion and path planning; AUTONOMOUS EXPLORATION; EFFICIENT;
D O I
10.1109/LRA.2024.3379804
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In this work, we propose a deep reinforcement learning (DRL) based reactive planner to solve large-scale Lidar-based autonomous robot exploration problems in 2D action space. Our DRL-based planner allows the agent to reactively plan its exploration path by making implicit predictions about unknown areas, based on a learned estimation of the underlying transition model of the environment. To this end, our approach relies on learned attention mechanisms for their powerful ability to capture long-term dependencies at different spatial scales to reason about the robot's entire belief over known areas. Our approach relies on ground truth information (i.e., privileged learning) to guide the environment estimation during training, as well as on a graph rarefaction algorithm, which allows models trained in small-scale environments to scale to large-scale ones. Simulation results show that our model exhibits better exploration efficiency (12% in path length, 6% in makespan) and lower planning time (60%) than the state-of-the-art planners in a 130 m x 100 m benchmark scenario. We also validate our learned model on hardware.
引用
收藏
页码:4631 / 4638
页数:8
相关论文
共 34 条
  • [1] Semantic OcTree Mapping and Shannon Mutual Information Computation for Robot Exploration
    Asgharivaskasi, Arash
    Atanasov, Nikolay
    [J]. IEEE TRANSACTIONS ON ROBOTICS, 2023, 39 (03) : 1910 - 1928
  • [2] Active Mapping via Gradient Ascent Optimization of Shannon Mutual Information over Continuous SE(3) Trajectories
    Asgharivaskasi, Arash
    Koga, Shumon
    Atanasov, Nikolay
    [J]. 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 12994 - 13001
  • [3] Bircher A, 2016, IEEE INT CONF ROBOT, P1462, DOI 10.1109/ICRA.2016.7487281
  • [4] Representation granularity enables time-efficient autonomous exploration in large, complex worlds
    Cao, C.
    Zhu, H.
    Ren, Z.
    Choset, H.
    Zhang, J.
    [J]. SCIENCE ROBOTICS, 2023, 8 (80)
  • [5] Cao C, 2021, ROBOT SCI SYS
  • [6] Cao Y., 2023, Conference on Robot Learning. PMLR, P1928
  • [7] Cao Y., 2022, P INT S DISTR AUT RO, P202
  • [8] ARiADNE: A Reinforcement learning approach using Attention-based Deep Networks for Exploration
    Cao, Yuhong
    Hou, Tianxiang
    Wang, Yizhuo
    Yi, Xian
    Sartoretti, Guillaume
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 10219 - 10225
  • [9] Chen F., 2019, P AIAA SCIT FOR
  • [10] Autonomous Exploration Under Uncertainty via Deep Reinforcement Learning on Graphs
    Chen, Fanfei
    Martin, John D.
    Huang, Yewei
    Wang, Jinkun
    Englot, Brendan
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 6140 - 6147