Robot Dynamic Path Planning Based on Prioritized Experience Replay and LSTM Network

被引:0
|
作者
Li, Hongqi [1 ]
Zhong, Peisi [1 ]
Liu, Li [2 ]
Wang, Xiao [1 ]
Liu, Mei [3 ]
Yuan, Jie [1 ]
机构
[1] Shandong Univ Sci & Technol, Coll Mech & Elect Engn, Qingdao 266590, Peoples R China
[2] Shandong Univ Sci & Technol, Coll Comp Sci & Engn, Qingdao 266590, Peoples R China
[3] Shandong Univ Sci & Technol, Coll Energy Storage Technol, Qingdao 266590, Peoples R China
来源
IEEE ACCESS | 2025年 / 13卷
基金
中国国家自然科学基金;
关键词
Heuristic algorithms; Long short term memory; Path planning; Convergence; Robots; Training; Planning; Adaptation models; Accuracy; Deep reinforcement learning; DDQN; LSTM network; mobile robot; path planning; prioritized experience replay; LEARNING ALGORITHM;
D O I
10.1109/ACCESS.2025.3532449
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To address the issues of slow convergence speed, poor dynamic adaptability, and path redundancy in the Double Deep Q Network (DDQN) within complex obstacle environments, this paper proposes an enhanced algorithm within the deep reinforcement learning framework. This algorithm, termed LPDDQN, integrates Prioritized Experience Replay (PER) and the Long Short Term Memory (LSTM) network to improve upon the DDQN algorithm. First, Prioritized Experience Replay (PER) is utilized to prioritize experience data and optimize storage and sampling operations through the SumTree structure, rather than the conventional experience queue. Second, the LSTM network is introduced to enhance the dynamic adaptability of the DDQN algorithm. Owing to the introduction of the LSTM model, the experience samples must be sliced and populated. The performance of the proposed LPDDQN algorithm is compared with five other path planning algorithms in both static and dynamic environments. Simulation analysis shows that in a static environment, LPDDQN demonstrates significant improvements over traditional DDQN in terms of convergence, number of moving steps, success rate, and number of turns, with respective improvements of 24.07%, 17.49%, 37.73%, and 61.54%. In dynamic and complex environments, the success rates of all algorithms, except TLD3 and the LPDDQN, decreased significantly. Further analysis reveals that the LPDDQN outperforms the TLD3 by 18.87%, 2.41%, and 39.02% in terms of moving steps, success rate, and number of turns, respectively.
引用
收藏
页码:22283 / 22299
页数:17
相关论文
共 50 条
  • [41] Robot dynamic path planning based on improved ant colony and DWA algorithm
    Wei L.-X.
    Zhang Y.-K.
    Sun H.
    Hou S.-J.
    Kongzhi yu Juece/Control and Decision, 2022, 37 (09): : 2211 - 2216
  • [42] Research on dynamic path planning method of electric inspection robot based on fuzzy neural network
    Jiang, Yang
    Yu, Yifan
    Sun, Ruohuai
    Zhao, Bin
    ENERGY REPORTS, 2023, 9 : 483 - 491
  • [43] Research on dynamic path planning method of electric inspection robot based on fuzzy neural network
    Jiang, Yang
    Yu, Yifan
    Sun, Ruohuai
    Zhao, Bin
    ENERGY REPORTS, 2023, 9 : 483 - 491
  • [44] Asynchronous Deep Q-network in Continuous Environment Based on Prioritized Experience Replay
    Liu, Hongda
    Zhang, Hanqi
    Gong, Linying
    2019 2ND INTERNATIONAL CONFERENCE ON MECHANICAL ENGINEERING, INDUSTRIAL MATERIALS AND INDUSTRIAL ELECTRONICS (MEIMIE 2019), 2019, : 472 - 477
  • [45] Mobile Robot Path Planning Based on Improved A* Algorithm and Dynamic Window Method
    Wang H.
    Yin P.
    Zheng W.
    Wang H.
    Zuo J.
    1600, Chinese Academy of Sciences (42): : 346 - 353
  • [46] Research on Mobile Robot Path Planning in Dynamic Fnvironment
    Liang, Xianxia
    Liu, Chaoying
    Song, Xueling
    Hao, Cunming
    2017 CHINESE AUTOMATION CONGRESS (CAC), 2017, : 3890 - 3894
  • [47] A new path-planning algorithm for mobile robot based on neural network
    Zhu, YJ
    Chang, J
    Wang, SG
    2002 IEEE REGION 10 CONFERENCE ON COMPUTERS, COMMUNICATIONS, CONTROL AND POWER ENGINEERING, VOLS I-III, PROCEEDINGS, 2002, : 1570 - 1573
  • [48] Path Planning for Nonholonomic Mobile Robot in Dynamic Environment
    Kang Liang
    Mao Lian-cheng
    PROGRESS IN MECHATRONICS AND INFORMATION TECHNOLOGY, PTS 1 AND 2, 2014, 462-463 : 771 - 774
  • [49] A Multi-Scale Map Method Based on Bioinspired Neural Network Algorithm for Robot Path Planning
    Luo, Min
    Hou, Xiaorong
    Yang, Simon X.
    IEEE ACCESS, 2019, 7 : 142682 - 142691
  • [50] Research on path planning of robot based on deep reinforcement learning
    Liu, Feng
    Chen, Chang
    Li, Zhihua
    Guan, Zhi-Hong
    Wang, Hua O.
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 3730 - 3734