Random Network Distillation Based Deep Reinforcement Learning for AGV Path Planning

被引:1
|
作者
Yin, Huilin [1 ]
Su, Shengkai [1 ]
Lin, Yinjia [1 ]
Zhen, Pengju [1 ]
Festl, Karin [2 ]
Watzenig, Daniel [2 ]
机构
[1] Tongji Univ, Sch Elect & Informat Engn, Shanghai, Peoples R China
[2] Graz Univ Technol, Virtual Vehicle Res GmbH, A-8010 Graz, Austria
基金
中国国家自然科学基金;
关键词
D O I
10.1109/IV55156.2024.10588651
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the flourishing development of intelligent warehousing systems, the technology of Automated Guided Vehicle (AGV) has experienced rapid growth. Within intelligent warehousing environments, AGV is required to safely and rapidly plan an optimal path in complex and dynamic environments. Most research has studied deep reinforcement learning to address this challenge. However, in the environments with sparse extrinsic rewards, these algorithms often converge slowly, learn inefficiently or fail to reach the target. Random Network Distillation (RND), as an exploration enhancement, can effectively improve the performance of proximal policy optimization, especially enhancing the additional intrinsic rewards of the AGV agent which is in sparse reward environments. Moreover, most of the current research continues to use 2D grid mazes as experimental environments. These environments have insufficient complexity and limited action sets. To solve this limitation, we present simulation environments of AGV path planning with continuous actions and positions for AGVs, so that it can be close to realistic physical scenarios. Based on our experiments and comprehensive analysis of the proposed method, the results demonstrate that our proposed method enables AGV to more rapidly complete path planning tasks with continuous actions in our environments. A video of part of our experiments can be found at https://youtu.be/lwrY9YesGmw.
引用
收藏
页码:2667 / 2673
页数:7
相关论文
共 50 条
  • [1] AGV Path Planning Model based on Reinforcement Learning
    Liao, Xiaofei
    Wang, Yang
    Xuan, Yiliang
    Wu, Dequan
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 6722 - 6726
  • [2] AGV Path Planning Based on Memristor Reinforcement Learning in Warehouse Environment
    Yang, Hailan
    Qi, Yongqiang
    Rong, Dan
    Computer Engineering and Applications, 2023, 59 (17) : 318 - 327
  • [3] Multi-AGV motion planning based on deep reinforcement learning
    Sun H.
    Yuan W.
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2024, 30 (02): : 708 - 716
  • [4] Robot path planning based on deep reinforcement learning
    Long, Yinxin
    He, Huajin
    2020 IEEE CONFERENCE ON TELECOMMUNICATIONS, OPTICS AND COMPUTER SCIENCE (TOCS), 2020, : 151 - 154
  • [5] Robot Path Planning Based on Deep Reinforcement Learning
    Zhang, Rui
    Jiang, Yuhao
    Wu Fenghua
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 1697 - 1701
  • [6] A Multi-AGV Routing Planning Method Based on Deep Reinforcement Learning and Recurrent Neural Network
    Yishuai Lin
    Gang Hu
    Liang Wang
    Qingshan Li
    Jiawei Zhu
    IEEE/CAA Journal of Automatica Sinica, 2024, 11 (07) : 1720 - 1722
  • [7] A Multi-AGV Routing Planning Method Based on Deep Reinforcement Learning and Recurrent Neural Network
    Lin, Yishuai
    Hue, Gang
    Wang, Liang
    Li, Qingshan
    Zhu, Jiawei
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2024, 11 (07) : 1720 - 1722
  • [8] Learning Navigation Policies for Mobile Robots in Deep Reinforcement Learning with Random Network Distillation
    Pan, Lifan
    Li, Anyi
    Ma, Jun
    Ji, Jianmin
    2021 5TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE (ICIAI 2021), 2021, : 151 - 157
  • [9] Reinforcement learning method based on sample regularization and adaptive learning rate for AGV path planning
    Nie, Jun
    Zhang, Guihua
    Lu, Xiao
    Wang, Haixia
    Sheng, Chunyang
    Sun, Lijie
    NEUROCOMPUTING, 2025, 614
  • [10] UCAV Path Planning Algorithm Based on Deep Reinforcement Learning
    Zheng, Kaiyuan
    Gao, Jingpeng
    Shen, Liangxi
    IMAGE AND GRAPHICS, ICIG 2019, PT II, 2019, 11902 : 702 - 714