A path planning strategy unified with a COLREGS collision avoidance function based on deep reinforcement learning and artificial potential field

被引:182
作者
Li, Lingyu [1 ]
Wu, Defeng [1 ,2 ]
Huang, Youqiang [1 ]
Yuan, Zhi-Ming [3 ]
机构
[1] Jimei Univ, Sch Marine Engn, Xiamen 361021, Peoples R China
[2] Fujian Prov Key Lab Naval Architecture & Ocean En, Xiamen 361021, Peoples R China
[3] Univ Strathclyde, Dept Naval Architecture Ocean & Marine Engn, Glasgow G4 0LZ, Lanark, Scotland
基金
中国国家自然科学基金;
关键词
Deep reinforcement learning; Path planning; Artificial potential field; COLREGS collision avoidance;
D O I
10.1016/j.apor.2021.102759
中图分类号
P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
Improving the autopilot capability of ships is particularly important to ensure the safety of maritime navigation. The unmanned surface vessel (USV) with autopilot capability is a development trend of the ship of the future. The objective of this paper is to investigate the path planning problem of USVs in uncertain environments, and a path planning strategy unified with a collision avoidance function based on deep reinforcement learning (DRL) is proposed. A Deep Q-learning network (DQN) is used to continuously interact with the visually simulated environment to obtain experience data, so that the agent learns the best action strategies in the visual simulated environment. To solve the collision avoidance problems that may occur during USV navigation, the location of the obstacle ship is divided into four collision avoidance zones according to the International Regulations for Preventing Collisions at Sea (COLREGS). To obtain an improved DRL algorithm, the artificial potential field (APF) algorithm is utilized to improve the action space and reward function of the DQN algorithm. A simulation experiments is utilized to test the effects of our method in various situations. It is also shown that the enhanced DRL can effectively realize autonomous collision avoidance path planning.
引用
收藏
页数:16
相关论文
共 39 条
[1]  
[Anonymous], 2016, INT C LEARN REPR
[2]   Deep learning [J].
LeCun, Yann ;
Bengio, Yoshua ;
Hinton, Geoffrey .
NATURE, 2015, 521 (7553) :436-444
[3]  
Beser Fuat, 2018, Procedia Computer Science, V131, P633, DOI 10.1016/j.procs.2018.04.306
[4]  
Campbell S., 2012, IFAC P VOLUMES IFAC, P386, DOI DOI 10.3182/20120919-3-IT-2046.00066
[5]   A knowledge-free path planning approach for smart ships based on reinforcement learning [J].
Chen, Chen ;
Chen, Xian-Qiao ;
Ma, Feng ;
Zeng, Xiao-Jun ;
Wang, Jin .
OCEAN ENGINEERING, 2019, 189
[6]   MARINE TRAFFIC BEHAVIOR IN RESTRICTED WATERS [J].
COLDWELL, TG .
JOURNAL OF NAVIGATION, 1983, 36 (03) :430-444
[7]  
Dann M, 2019, AAAI CONF ARTIF INTE, P881
[8]  
Ding F., 2018, OCEANS 2018 MTS IEEE, P1
[9]   Neural networks based reinforcement learning for mobile robots obstacle avoidance [J].
Duguleana, Mihai ;
Mogan, Gheorghe .
EXPERT SYSTEMS WITH APPLICATIONS, 2016, 62 :104-115
[10]   STATISTICAL STUDY OF SHIP DOMAINS [J].
GOODWIN, EM .
JOURNAL OF NAVIGATION, 1975, 28 (03) :328-344