A path planning algorithm fusion of obstacle avoidance and memory functions

被引:0
作者
Zheng, Qingchun [1 ,2 ]
Li, Shubo [1 ,2 ]
Zhu, Peihao [1 ,2 ]
Ma, Wenpeng [1 ,2 ]
Wang, Yanlu [1 ,2 ]
机构
[1] Tianjin Univ Technol, Sch Mech Engn, Tianjin Key Lab Adv Mechatron Syst Design & Intell, Tianjin, Peoples R China
[2] Tianjin Univ Technol, Natl Demonstrat Ctr Expt Mech & Elect Engn Educ, Tianjin, Peoples R China
基金
中国国家自然科学基金;
关键词
artificial intelligence; deep reinforcement learning; intelligent robots; mobile robots; path planning; NAVIGATION;
D O I
10.1049/ccs2.12098
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, to address the issues of sluggish convergence and poor learning efficiency at the initial stages of training, the authors improve and optimise the Deep Deterministic Policy Gradient (DDPG) algorithm. First, inspired by the Artificial Potential Field method, the selection strategy of DDPG has been improved to accelerate the convergence speed during the early stages of training and reduce the time it takes for the mobile robot to reach the target point. Then, optimising the neural network structure of the DDPG algorithm based on the Long Short-Term Memory accelerates the algorithm's convergence speed in complex dynamic scenes. Static and dynamic scene simulation experiments of mobile robots are carried out in ROS. Test findings demonstrate that the Artificial Potential Field method-Long Short Term Memory Deep Deterministic Policy Gradient (APF-LSTM DDPG) algorithm converges significantly faster in complex dynamic scenes. The success rate is improved by 7.3% and 3.6% in contrast to the DDPG and LSTM-DDPG algorithms. Finally, the usefulness of the method provided in this study is similarly demonstrated in real situations using real mobile robot platforms, laying the foundation for the path planning of mobile robots in complex changing conditions.
引用
收藏
页码:300 / 313
页数:14
相关论文
共 22 条
[1]   Deep Deterministic Policy Gradient for Navigation of Mobile Robots [J].
de Jesus, Junior Costa ;
Bottega, Jair Augusto ;
de Souza Leite Cuadros, Marco Antonio ;
Tello Gamarra, Daniel Fernando .
JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (01) :349-361
[2]   The dynamic window approach to collision avoidance [J].
Fox, D ;
Burgard, W ;
Thrun, S .
IEEE ROBOTICS & AUTOMATION MAGAZINE, 1997, 4 (01) :23-33
[3]   Improved Deep Deterministic Policy Gradient for Dynamic Obstacle Avoidance of Mobile Robot [J].
Gao, Xiaoshan ;
Yan, Liang ;
Li, Zhijun ;
Wang, Gang ;
Chen, I-Ming .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (06) :3675-3682
[4]   Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient [J].
Gong, Hui ;
Wang, Peng ;
Ni, Cui ;
Cheng, Nuo .
SENSORS, 2022, 22 (09)
[6]   Predictive hierarchical reinforcement learning for path-efficient mapless navigation with moving target [J].
Li, Hanxiao ;
Luo, Biao ;
Song, Wei ;
Yang, Chunhua .
NEURAL NETWORKS, 2023, 165 :677-688
[7]  
Lillicrap T. P., 2015, ARXIV
[8]   Collision-free path planning for a guava-harvesting robot based on recurrent deep reinforcement learning [J].
Lin, Guichao ;
Zhu, Lixue ;
Li, Jinhui ;
Zou, Xiangjun ;
Tang, Yunchao .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2021, 188
[9]   A Multi-AGV Routing Planning Method Based on Deep Reinforcement Learning and Recurrent Neural Network [J].
Lin, Yishuai ;
Hue, Gang ;
Wang, Liang ;
Li, Qingshan ;
Zhu, Jiawei .
IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2024, 11 (07) :1720-1722
[10]   Visual Navigation for Biped Humanoid Robots Using Deep Reinforcement Learning [J].
Lobos-Tsunekawa, Kenzo ;
Leiva, Francisco ;
Ruiz-del-Solar, Javier .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :3247-3254