Research on path planning algorithm of mobile robot based on reinforcement learning

被引:20
作者
Pan, Guoqian [1 ]
Xiang, Yong [2 ,3 ]
Wang, Xiaorui [2 ,3 ]
Yu, Zhongquan [2 ,3 ]
Zhou, Xinzhi [1 ]
机构
[1] Sichuan Univ, Coll Elect & Informat Engn, Chengdu, Sichuan, Peoples R China
[2] CAAC, Res Inst 2, Chengdu, Sichuan, Peoples R China
[3] Civil Aviat Logist Technol Co Ltd, Chengdu, Sichuan, Peoples R China
基金
中国国家自然科学基金;
关键词
Complex environment; Mobile robot; Path planning; Q-learning algorithm; NAVIGATION;
D O I
10.1007/s00500-022-07293-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In order to solve the problems of low learning efficiency and slow convergence speed when mobile robot uses reinforcement learning method for path planning in complex environment, a reinforcement learning method based on each round path planning result is proposed. Firstly, the algorithm adds obstacle learning matrix to improve the success rate of path planning; and introduces heuristic reward to speed up the learning process by reducing the search space; then proposes a method of dynamically adjusting the exploration factor to balance the exploration and utilization in path planning, so as to further improve the performance of the algorithm. Finally, the simulation experiment in grid environment shows that compared with Q-learning algorithm, the improved algorithm not only shortens the average path length of the robot to reach the target position, but also speeds up the learning efficiency of the algorithm, so that the robot can find the optimal path more quickly. The code of EPRQL algorithm proposed in this paper has been published to GitHub: GitHub: https://github.com/ panpanpanguoguoqian/ mypaper1.git.
引用
收藏
页码:8961 / 8970
页数:10
相关论文
共 18 条
[1]   Multiple objective genetic algorithms for path-planning optimization in autonomous mobile robots [J].
Castillo, Oscar ;
Trujillo, Leonardo ;
Melin, Patricia .
SOFT COMPUTING, 2007, 11 (03) :269-279
[2]   A new Q-learning algorithm based on the Metropolis criterion [J].
Guo, MZ ;
Liu, Y ;
Malec, J .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2004, 34 (05) :2140-2143
[3]   Reinforcement based mobile robot navigation in dynamic environment [J].
Jaradat, Mohammad Abdel Kareem ;
Al-Rousan, Mohammad ;
Quadan, Lara .
ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2011, 27 (01) :135-149
[4]   Supervised Neural Q_learning based Motion Control for Bionic Underwater Robots [J].
Lin, Longxin ;
Xie, Haibin ;
Zhang, Daibing ;
Shen, Lincheng .
JOURNAL OF BIONIC ENGINEERING, 2010, 7 :S177-S184
[5]   Solving the optimal path planning of a mobile robot using improved Q-learning [J].
Low, Ee Soong ;
Ong, Pauline ;
Cheah, Kah Chun .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 115 :143-161
[6]   Path Planning for Autonomous Underwater Vehicles: An Ant Colony Algorithm Incorporating Alarm Pheromone [J].
Ma, Yi-Ning ;
Gong, Yue-Jiao ;
Xiao, Chu-Feng ;
Gao, Ying ;
Zhang, Jun .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (01) :141-154
[7]   Robot Reinforcement Learning for Automatically Avoiding a Dynamic Obstacle in a Virtual Environment [J].
Phuong Chu ;
Hoang Vu ;
Yeo, Donghyeon ;
Lee, Byeonggwon ;
Um, Kyhyun ;
Cho, Kyungeun .
ADVANCED MULTIMEDIA AND UBIQUITOUS ENGINEERING: FUTURE INFORMATION TECHNOLOGY, 2015, 352 :157-164
[8]   Survey of Model-Based Reinforcement Learning: Applications on Robotics [J].
Polydoros, Athanasios S. ;
Nalpantidis, Lazaros .
JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2017, 86 (02) :153-173
[9]   EXACT ROBOT NAVIGATION USING ARTIFICIAL POTENTIAL FUNCTIONS [J].
RIMON, E ;
KODITSCHEK, DE .
IEEE TRANSACTIONS ON ROBOTICS AND AUTOMATION, 1992, 8 (05) :501-518
[10]  
Song Yong, 2012, Control Theory & Applications, V29, P1623