Research on obstacle avoidance of underactuated autonomous underwater vehicle based on offline reinforcement learning

被引:0
作者
Liu, Tao [1 ,2 ]
Huang, Junhao [1 ]
Zhao, Jintao [1 ]
机构
[1] Sun Yat Sen Univ & Southern Marine Sci & Engn, Sch Ocean Engn & Technol, Guangdong Lab Zhuhai, Zhuhai, Peoples R China
[2] Guangdong Prov Key Lab Informat Technol Deep Water, Zhuhai, Peoples R China
关键词
AUVs; offline rL; CQL; dynamic obstacle avoidance; AUV;
D O I
10.1017/S0263574724001802
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
The autonomous navigation and obstacle avoidance capabilities of autonomous underwater vehicles (AUVs) are essential for ensuring their safe navigation and long-term, efficient operation. However, the complexity of the marine environment poses significant challenges to safe and effective obstacle avoidance. To address this issue, this study proposes an AUV obstacle avoidance control algorithm based on offline reinforcement learning. This method adopts the Conservative Q-learning (CQL) algorithm, which is based on the Soft Actor-Critic (SAC) framework. It learns from obtained historical obstacle avoidance data and ultimately achieves a favorable obstacle avoidance control strategy. In this method, PID and SAC control algorithms are utilized to generate expert obstacle avoidance data to construct a diversified offline database. Additionally, based on the line-of-sight (LOS) guidance method and artificial potential field (APF) method, information regarding the distance and orientation of targets and obstacles is incorporated into the state space, and heading and obstacle avoidance reward terms are integrated into the reward functiondesign. The algorithm successfully guides the AUV in autonomous navigation and dynamic obstacle avoidance inthree-dimensional space. Furthermore, the algorithm exhibits a certain degree of anti-interference capability against uncertain disturbances and ocean currents, enhancing the safety and robustness of the AUV system. Simulation results fully demonstrate the feasibility and effectiveness of the intelligent obstacle avoidance method based on offline reinforcement learning. This study highlights the profound significance of offline reinforcement learning in enabling robust and reliable control systems for AUVs, paving the way for enhanced operational capabilities inchallenging marine environments
引用
收藏
页码:194 / 218
页数:25
相关论文
共 36 条
[1]  
Agarwal R, 2020, PR MACH LEARN RES, V119
[2]   Adaptive low-level control of autonomous underwater vehicles using deep reinforcement learning [J].
Carlucho, Ignacio ;
De Paula, Mariano ;
Wang, Sen ;
Petillot, Yvan ;
Acosta, Gerardo G. .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2018, 107 :71-86
[3]   Path planning and obstacle avoidance for AUV: A review [J].
Cheng, Chunxi ;
Sha, Qixin ;
He, Bo ;
Li, Guangliang .
OCEAN ENGINEERING, 2021, 235
[4]   Adaptive Neural Network Control of AUVs With Control Input Nonlinearities Using Reinforcement Learning [J].
Cui, Rongxin ;
Yang, Chenguang ;
Li, Yang ;
Sharma, Sanjay .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2017, 47 (06) :1019-1029
[5]   AUV position tracking and trajectory control based on fast-deployed deep reinforcement learning method [J].
Fang, Yuan ;
Huang, Zhenwei ;
Pu, Jinyun ;
Zhang, Jinsong .
OCEAN ENGINEERING, 2022, 245
[6]  
Fossen T.I., 2011, HDB MARINE CRAFT HYD, DOI DOI 10.1002/9781119994138
[7]  
Fossen T.I., 2002, MARINE CONTROL SYSTE
[8]  
Fujimoto S, 2019, PR MACH LEARN RES, V97
[9]   Trajectory tracking control for autonomous underwater vehicles based on dual closed-loop of MPC with uncertain dynamics [J].
Gong, Peng ;
Yan, Zheping ;
Zhang, Wei ;
Tang, Jialing .
OCEAN ENGINEERING, 2022, 265
[10]   Deep reinforcement learning for adaptive path planning and control of an autonomous underwater vehicle [J].
Hadi, Behnaz ;
Khosravi, Alireza ;
Sarhadi, Pouria .
APPLIED OCEAN RESEARCH, 2022, 129