Path Planning for Mobile Robots Using Transfer Reinforcement Learning

被引:2
作者
Zheng, Xinwang [1 ]
Zheng, Wenjie [2 ]
Du, Yong [2 ]
Li, Tiejun [2 ]
Yuan, Zhansheng [2 ]
机构
[1] Jimei Univ, Chengyi Coll, Xiamen 361021, Fujian, Peoples R China
[2] Jimei Univ, Sch Ocean Informat Engn, Xiamen 361021, Fujian, Peoples R China
关键词
Deep reinforcement transfer learning; heterogeneous environment; path planning; robot navigation; NAVIGATION; STRATEGIES;
D O I
10.1142/S0218213024400050
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The path planning of mobile robots helps robots to perceive environment using the information obtained from sensors and plan a route to reach the target. With the increasing difficulty of task, the environment the mobile robots face becomes more and more complex. Traditional path planning methods can no longer meet the requirements of mobile robot navigation in complex environment. Deep reinforcement learning (DRL) is introduced into robot navigation However, it may be time-consuming to train DRL model when the environment is very complex and the existing environment may differ from the unknown environment. In order to handle the robot navigation in heterogeneous environment, this paper utilizes deep transfer reinforcement learning (DTRL) for mobile robot path planning. Compared with DRL, DTRL does not require the distribution of the existing environment is the same as that of the unknown environment. Additionally, DTRL can transfer the knowledge of existing model to new scenario to reduce the training time. The simulations show that DTRL can reach higher success rate than DRL for heterogeneous environment robot navigation. By using local policy, it costs less time to train DTRL than DRL for a complex environment and DTRL can consume less navigation time.
引用
收藏
页数:15
相关论文
共 48 条
[21]   Path planning techniques for mobile robots: Review and prospect [J].
Liu, Lixing ;
Wang, Xu ;
Yang, Xin ;
Liu, Hongjie ;
Li, Jianping ;
Wang, Pengfei .
EXPERT SYSTEMS WITH APPLICATIONS, 2023, 227
[22]  
Liu T., 2014, MEASUREMENT, V2023
[23]   Deep learning-based welding image recognition: A comprehensive review [J].
Liu, Tianyuan ;
Zheng, Pai ;
Bao, Jinsong .
JOURNAL OF MANUFACTURING SYSTEMS, 2023, 68 :601-625
[24]   Domain adaptive learning based on equilibrium distribution and dynamic subspace approximation [J].
Liu, Zhonghua ;
Wang, Tiansheng ;
Zhu, Fa ;
Chen, Xingchi ;
Pelusi, Danilo ;
Vasilakos, Athanasios V. .
EXPERT SYSTEMS WITH APPLICATIONS, 2024, 249
[25]   Manifold transfer subspace learning based on double relaxed discriminative regression [J].
Liu, Zhonghua ;
Zhu, Fa ;
Zhang, Kaibing ;
Lai, Zhihui ;
Huo, Hua .
ARTIFICIAL INTELLIGENCE REVIEW, 2023, 56 (SUPPL 1) :959-981
[26]   Instance-Level Knowledge Transfer for Data-Driven Driver Model Adaptation With Homogeneous Domains [J].
Lu, Chao ;
Lv, Chen ;
Gong, Jianwei ;
Wang, Wenshuo ;
Cao, Dongpu ;
Wang, Fei-Yue .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (10) :17015-17026
[27]   PATH-PLANNING STRATEGIES FOR A POINT MOBILE AUTOMATON MOVING AMIDST UNKNOWN OBSTACLES OF ARBITRARY SHAPE [J].
LUMELSKY, VJ ;
STEPANOV, AA .
ALGORITHMICA, 1987, 2 (04) :403-430
[28]   Path planning of multi-agent systems in unknown environment with neural kernel smoothing and reinforcement learning [J].
Luviano Cruz, David ;
Yu, Wen .
NEUROCOMPUTING, 2017, 233 :34-42
[29]  
Mnih V., ARXIV
[30]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533