Costmap-based Local Motion Planning using Deep Reinforcement Learning

被引:0
|
作者
Garrote, Luis [1 ,2 ]
Perdiz, Joao [1 ,2 ]
Nunes, Urbano J. [1 ,2 ]
机构
[1] Univ Coimbra, Dept Elect & Comp Engn, Coimbra, Portugal
[2] Univ Coimbra, Inst Syst & Robot, Coimbra, Portugal
关键词
D O I
10.1109/RO-MAN57019.2023.10309389
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Local motion planning is an essential component of autonomous robot navigation systems as it involves generating collision-free trajectories for the robot in real-time, given its current position, the map of the environment and a goal. Considering an a priori goal path, computed by a global planner or as the output of a mission planning approach, this paper proposes a Two-Stream Deep Reinforcement Learning strategy for local motion planning that takes as inputs a local costmap representing the robot's surrounding obstacles and a local costmap representing the nearest goal path. The proposed approach uses a Double Dueling Deep Q-Network and a new reward model to avoid obstacles while trying to maintain the lateral error between the robot and the goal path close to zero. Our approach enables the robot to navigate through complex environments, including cluttered spaces and narrow passages, while avoiding collisions with obstacles. Evaluation of the proposed approach was carried out in an in-house simulation environment, in five scenarios. Double and Double Dueling architectures were evaluated; the presented results show that the proposed strategy can correctly follow the desired goal path and, when needed, avoid obstacles ahead and recover back to following the goal path.
引用
收藏
页码:1089 / 1095
页数:7
相关论文
共 50 条
  • [41] Robot Path Planning Based on Deep Reinforcement Learning
    Zhang, Rui
    Jiang, Yuhao
    Wu Fenghua
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 1697 - 1701
  • [42] Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning
    You, Changxi
    Lu, Jianbo
    Filev, Dimitar
    Tsiotras, Panagiotis
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 114 : 1 - 18
  • [43] Real-time local path planning strategy based on deep distributional reinforcement learning
    Du, Shengli
    Zhu, Zexing
    Wang, Xuefang
    Han, Honggui
    Qiao, Junfei
    NEUROCOMPUTING, 2024, 599
  • [44] Deep reinforcement learning-based local path planning in dynamic environments for mobile robot☆
    Tao, Bodong
    Kim, Jae-Hoon
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2024, 36 (10)
  • [45] Harmonic-Based Optimal Motion Planning in Constrained Workspaces Using Reinforcement Learning
    Rousseas, Panagiotis
    Bechlioulis, Charalampos
    Kyriakopoulos, Kostas J.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 2005 - 2011
  • [46] Robotic Arm Motion Planning Based on Residual Reinforcement Learning
    Zhou, Dongxu
    Jia, Ruiqing
    Yao, Haifeng
    Xie, Mingzuo
    2021 THE 13TH INTERNATIONAL CONFERENCE ON COMPUTER AND AUTOMATION ENGINEERING (ICCAE 2021), 2021, : 89 - 94
  • [47] Humanoid motion planning of robotic arm based on reinforcement learning
    Yang A.
    Chen Y.
    Xu Y.
    Yi Qi Yi Biao Xue Bao/Chinese Journal of Scientific Instrument, 2021, 42 (12): : 136 - 145
  • [48] Deep Reinforcement Learning for Multi-contact Motion Planning of Hexapod Robots
    Fu, Huiqiao
    Tang, Kaiqiang
    Li, Peng
    Zhang, Wenqi
    Wang, Xinpeng
    Deng, Guizhou
    Wang, Tao
    Chen, Chunlin
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2381 - 2388
  • [49] Artificial Potential Field-Enhanced Deep Reinforcement Learning for Motion Planning
    Wang, Tengyue
    Chen, Xianhao
    Lei, Chen
    Yang, Liangjing
    2024 8TH INTERNATIONAL CONFERENCE ON ROBOTICS, CONTROL AND AUTOMATION, ICRCA 2024, 2024, : 459 - 465
  • [50] A Survey of Deep Reinforcement Learning Algorithms for Motion Planning and Control of Autonomous Vehicles
    Ye, Fei
    Zhang, Shen
    Wang, Pin
    Chan, Ching-Yao
    2021 32ND IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2021, : 1073 - 1080