Path Planning for Mobile Robots Using Transfer Reinforcement Learning

被引:2
作者
Zheng, Xinwang [1 ]
Zheng, Wenjie [2 ]
Du, Yong [2 ]
Li, Tiejun [2 ]
Yuan, Zhansheng [2 ]
机构
[1] Jimei Univ, Chengyi Coll, Xiamen 361021, Fujian, Peoples R China
[2] Jimei Univ, Sch Ocean Informat Engn, Xiamen 361021, Fujian, Peoples R China
关键词
Deep reinforcement transfer learning; heterogeneous environment; path planning; robot navigation; NAVIGATION; STRATEGIES;
D O I
10.1142/S0218213024400050
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The path planning of mobile robots helps robots to perceive environment using the information obtained from sensors and plan a route to reach the target. With the increasing difficulty of task, the environment the mobile robots face becomes more and more complex. Traditional path planning methods can no longer meet the requirements of mobile robot navigation in complex environment. Deep reinforcement learning (DRL) is introduced into robot navigation However, it may be time-consuming to train DRL model when the environment is very complex and the existing environment may differ from the unknown environment. In order to handle the robot navigation in heterogeneous environment, this paper utilizes deep transfer reinforcement learning (DTRL) for mobile robot path planning. Compared with DRL, DTRL does not require the distribution of the existing environment is the same as that of the unknown environment. Additionally, DTRL can transfer the knowledge of existing model to new scenario to reduce the training time. The simulations show that DTRL can reach higher success rate than DRL for heterogeneous environment robot navigation. By using local policy, it costs less time to train DTRL than DRL for a complex environment and DTRL can consume less navigation time.
引用
收藏
页数:15
相关论文
共 48 条
[31]  
Nie ZB, 2016, IEEE C EVOL COMPUTAT, P2532, DOI 10.1109/CEC.2016.7744104
[32]  
Okamura AM, 2010, IEEE ROBOT AUTOM MAG, V17, P26, DOI 10.1109/MRA.2010.937861
[33]   Robotic manipulation of multiple objects as a POMDP [J].
Pajarinen, Joni ;
Kyrki, Ville .
ARTIFICIAL INTELLIGENCE, 2017, 247 :213-228
[34]   A review: On path planning strategies for navigation of mobile robot [J].
Patle, B. K. ;
Babu, Ganesh L. ;
Pandey, Anish ;
Parhi, D. R. K. ;
Jagadeesh, A. .
DEFENCE TECHNOLOGY, 2019, 15 (04) :582-606
[35]   Motion planning and scheduling for human and industrial-robot collaboration [J].
Pellegrinelli, Stefania ;
Orlandini, Andrea ;
Pedrocchi, Nicola ;
Umbrico, Alessandro ;
Tolio, Tullio .
CIRP ANNALS-MANUFACTURING TECHNOLOGY, 2017, 66 (01) :1-4
[36]  
Pfeiffer Mark, 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA), P1527, DOI 10.1109/ICRA.2017.7989182
[37]  
Raja P., 2012, Int. J. Phys. Sci, V7, P1314, DOI DOI 10.5897/IJPS11.1745
[38]   Using genetic algorithm for drawing path planning in a robotic arm pencil sketching system [J].
Shaw, Jin-Siang ;
Lee, Szu-Ying .
PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART C-JOURNAL OF MECHANICAL ENGINEERING SCIENCE, 2024, 238 (14) :7134-7142
[39]   Route Planning for an Autonomous Robotic Vehicle Employing a Weight-Controlled Particle Swarm-Optimized Dijkstra Algorithm [J].
Sundarraj, Subaselvi ;
Reddy, R. Vijaya Kumar ;
Basam, Mahesh Babu ;
Lokesh, Gururaj Harinahalli ;
Flammini, Francesco ;
Natarajan, Rajesh .
IEEE ACCESS, 2023, 11 :92433-92442
[40]  
Tai L., ARXIV