Mapless navigation via Hierarchical Reinforcement Learning with memory-decaying novelty

被引:0
作者
Gao, Yan [1 ]
Lin, Feiqiang [1 ]
Cai, Boliang [1 ]
Wu, Jing [2 ]
Wei, Changyun [3 ]
Grech, Raphael [4 ]
Ji, Ze [1 ]
机构
[1] Cardiff Univ, Sch Engn, Cardiff CF24 3AA, Wales
[2] Cardiff Univ, Sch Comp Sci & Informat, Cardiff CF24 4AG, Wales
[3] Hohai Univ, Coll Mech & Elect Engn, Changzhou, Peoples R China
[4] Spirent Commun, Paignton TQ4 7QR, England
关键词
Mapless navigation; Deep reinforcement learning; Collision avoidance; Hierarchical Reinforcement Learning; Path planning; DEEP NEURAL-NETWORKS;
D O I
10.1016/j.robot.2024.104815
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Hierarchical Reinforcement Learning (HRL) has shown superior performance for mapless navigation tasks. However, it remains limited in unstructured environments that might contain terrains like long corridors and dead corners, which can lead to local minima. This is because most HRL-based mapless navigation methods employ a simplified reward setting and exploration strategy. In this work, we propose a novel reward function for training the high-level (HL) policy, which contains two components: extrinsic reward and intrinsic reward. The extrinsic reward encourages the robot to move towards the target location, while the intrinsic reward is computed based on novelty, episode memory and memory decaying, making the agent capable of accomplishing spontaneous exploration. We also design a novel neural network structure that incorporates an LSTM network to augment the agent with memory and reasoning capabilities. We test our method in unknown environments and specific scenarios prone to the local minimum problem to evaluate the navigation performance and local minimum resolution ability. The results show that our method significantly increases the success rate when compared to advanced RL-based methods, achieving a maximum improvement of nearly 28%. Our method demonstrates effective improvement in addressing the local minimum issue, especially in cases where the baselines fail completely. Additionally, numerous ablation studies consistently confirm the effectiveness of our proposed reward function and neural network structure.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Electric vehicle charging navigation method based on hierarchical reinforcement learning
    Zhan H.
    Jiang C.
    Su Q.
    [J]. Dianli Zidonghua Shebei/Electric Power Automation Equipment, 2022, 42 (10): : 264 - 272
  • [22] An Efficient Deep Reinforcement Learning Algorithm for Mapless Navigation with Gap-Guided Switching Strategy
    Li, Heng
    Qin, Jiahu
    Liu, Qingchen
    Yan, Chengzhen
    [J]. JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2023, 108 (03)
  • [23] Deep reinforcement learning based mapless navigation for industrial AMRs: advancements in generalization via potential risk state augmentation
    Xu, Degang
    Chen, Peng
    Zhou, Xianhan
    Wang, Yizhi
    Tan, Guanzheng
    [J]. APPLIED INTELLIGENCE, 2024, 54 (19) : 9295 - 9312
  • [24] Mapless Collaborative Navigation for a Multi-Robot System Based on the Deep Reinforcement Learning
    Chen, Wenzhou
    Zhou, Shizheng
    Pan, Zaisheng
    Zheng, Huixian
    Liu, Yong
    [J]. APPLIED SCIENCES-BASEL, 2019, 9 (20):
  • [25] Towards Efficient Mapless Navigation Using Deep Reinforcement Learning with Parameter Space Noise
    Liu, Xiaoyun
    Zhou, Qingrui
    Wang, Hui
    Yang, Ying
    [J]. PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 8833 - 8837
  • [26] Mapless Navigation with Deep Reinforcement Learning based on The Convolutional Proximal Policy Optimization Network
    Toan, Nguyen Duc
    Woo, Kim Gon
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP 2021), 2021, : 298 - 301
  • [27] Reinforced Imitation: Sample Efficient Deep Reinforcement Learning for Mapless Navigation by Leveraging Prior Demonstrations
    Pfeiffer, Mark
    Shukla, Samarth
    Turchetta, Matteo
    Cadena, Cesar
    Krause, Andreas
    Siegwart, Roland
    Nieto, Juan
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04): : 4423 - 4430
  • [28] Risk-aware deep reinforcement learning for mapless navigation of unmanned surface vehicles in uncertain and congested environments
    Wu, Xiangyu
    Wei, Changyun
    Guan, Dawei
    Ji, Ze
    [J]. OCEAN ENGINEERING, 2025, 322
  • [29] Hierarchical multi-robot navigation and formation in unknown environments via deep reinforcement learning and distributed optimization
    Chang, Lu
    Shan, Liang
    Zhang, Weilong
    Dai, Yuewei
    [J]. ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2023, 83
  • [30] Double Critic Deep Reinforcement Learning for Mapless 3D Navigation of Unmanned Aerial Vehicles
    Bedin Grando, Ricardo
    de Jesus, Junior Costa
    Kich, Victor Augusto
    Kolling, Alisson Henrique
    Jorge Drews-Jr, Paulo Lilles
    [J]. JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2022, 104 (02)