The Path Planning of Mobile Robot by Neural Networks and Hierarchical Reinforcement Learning

被引:93
作者
Yu, Jinglun [1 ]
Su, Yuancheng [1 ]
Liao, Yifan [1 ]
机构
[1] Chongqing Univ, Chongqing Univ Univ Cincinnati Joint Coop Inst, Chongqing, Peoples R China
关键词
neural network; hierarchical reinforcement learning; mobile robot; path planning; fusion algorithm; VIRTUAL-REALITY;
D O I
10.3389/fnbot.2020.00063
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing mobile robots cannot complete some functions. To solve these problems, which include autonomous learning in path planning, the slow convergence of path planning, and planned paths that are not smooth, it is possible to utilize neural networks to enable to the robot to perceive the environment and perform feature extraction, which enables them to have a fitness of environment to state action function. By mapping the current state of these actions through Hierarchical Reinforcement Learning (HRL), the needs of mobile robots are met. It is possible to construct a path planning model for mobile robots based on neural networks and HRL. In this article, the proposed algorithm is compared with different algorithms in path planning. It underwent a performance evaluation to obtain an optimal learning algorithm system. The optimal algorithm system was tested in different environments and scenarios to obtain optimal learning conditions, thereby verifying the effectiveness of the proposed algorithm. Deep Deterministic Policy Gradient (DDPG), a path planning algorithm for mobile robots based on neural networks and hierarchical reinforcement learning, performed better in all aspects than other algorithms. Specifically, when compared with Double Deep Q-Learning (DDQN), DDPG has a shorter path planning time and a reduced number of path steps. When introducing an influence value, this algorithm shortens the convergence time by 91% compared with the Q-learning algorithm and improves the smoothness of the planned path by 79%. The algorithm has a good generalization effect in different scenarios. These results have significance for research on guiding, the precise positioning, and path planning of mobile robots.
引用
收藏
页数:12
相关论文
共 34 条
[1]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[2]   Optimal path planning and execution for mobile robots using genetic algorithm and adaptive fuzzy-logic control [J].
Bakdi, Azzeddine ;
Hentout, Abdelfetah ;
Boutami, Hakim ;
Maoudj, Abderraouf ;
Hachour, Ouarda ;
Bouzouia, Brahim .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2017, 89 :95-109
[3]  
Botteghi N., 2020, ARXIV200204109, V2020, P1025
[4]   The research of human individual's conformity behavior in emergency situations [J].
Chen, Min .
LIBRARY HI TECH, 2020, 38 (03) :593-609
[5]   Analysis of FPA and BA meta-heuristic controllers for optimal path planning of mobile robot in cluttered environment [J].
Ghosh, Saradindu ;
Panigrahi, Pratap K. ;
Parhi, Dayal R. .
IET SCIENCE MEASUREMENT & TECHNOLOGY, 2017, 11 (07) :817-828
[6]   A Deep Q-Learning Approach for Dynamic Management of Heterogeneous Processors [J].
Gupta, Ujjwal ;
Mandal, Sumit K. ;
Mao, Manqing ;
Chakrabarti, Chaitali ;
Ogras, Umit Y. .
IEEE COMPUTER ARCHITECTURE LETTERS, 2019, 18 (01) :14-17
[7]  
Haarnoja T., 2018, ARXIV181205905, V2018, P26
[8]   Mobile robot path planning with surrounding point set and path improvement [J].
Han, Jihee ;
Seo, Yoonho .
APPLIED SOFT COMPUTING, 2017, 57 :35-47
[9]   Energy management based on reinforcement learning with double deep Q-learning for a hybrid electric tracked vehicle [J].
Han, Xuefeng ;
He, Hongwen ;
Wu, Jingda ;
Peng, Jiankun ;
Li, Yuecheng .
APPLIED ENERGY, 2019, 254
[10]   Dynamic Path Planning of Unknown Environment Based on Deep Reinforcement Learning [J].
Lei, Xiaoyun ;
Zhang, Zhian ;
Dong, Peifang .
JOURNAL OF ROBOTICS, 2018, 2018