Trajectory tracking control of vectored thruster autonomous underwater vehicles based on deep reinforcement learning

被引:0
作者
Liu, Tao [1 ,2 ]
Zhao, Jintao [1 ,2 ]
Hu, Yuli [3 ]
Huang, Junhao [1 ,2 ]
机构
[1] Sun Yat Sen Univ, Sch Ocean Engn & Technol, Zhuhai 519000, Peoples R China
[2] Southern Marine Sci & Engn Guangdong Lab Zhuhai, Zhuhai 519000, Peoples R China
[3] Northwestern Polytech Univ, Sch Marine Sci & Technol, Xian, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning (RL); autonomous underwater vehicle (AUV); trajectory tracking; extended state observer (ESO); environmental interference;
D O I
10.1080/17445302.2024.2391235
中图分类号
U6 [水路运输]; P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
Vectored thruster autonomous underwater vehicles (AUVs) offer superior manoeuvrability compared to traditional fin and rudder systems, especially at low or zero velocities. However, controlling these AUVs becomes challenging in the presence of unpredictable interference forces and external water flow velocities. This paper introduces a novel three-dimensional trajectory tracking method for vectored thruster AUVs using reinforcement learning, enabling the system to learn optimal control policies through environmental interaction. A 5-degree of freedom (5-DOF) model is developed from the kinematics and dynamics equations of the underactuated AUV. An extended state observer (ESO) is used to estimate the differential expansion state based on observed variables. A reinforcement learning-based trajectory tracking strategy is then implemented. Simulation experiments demonstrate the method's effectiveness in precisely controlling AUVs under varying environmental conditions, achieving exceptional tracking accuracy even in the presence of random interference forces and external water flow velocities.
引用
收藏
页数:14
相关论文
共 35 条
[1]   Continuous finite-time stabilization of the translational and rotational double integrators [J].
Bhat, SP ;
Bernstein, DS .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1998, 43 (05) :678-682
[2]  
Carlucho I, 2018, OCEANS 2018 MTS/IEEE CHARLESTON
[3]   A behavior-based scheme using reinforcement learning for autonomous underwater vehicles [J].
Carreras, M ;
Yuh, J ;
Batlle, J ;
Ridao, P .
IEEE JOURNAL OF OCEANIC ENGINEERING, 2005, 30 (02) :416-427
[4]   Path planning and obstacle avoidance for AUV: A review [J].
Cheng, Chunxi ;
Sha, Qixin ;
He, Bo ;
Li, Guangliang .
OCEAN ENGINEERING, 2021, 235
[5]   Horizontal Trajectory Tracking of Underactuated AUV using Backstepping Approach [J].
Cho, Gun Rae ;
Park, Dae-Gil ;
Kang, Hyungjoo ;
Lee, Mun-Jik ;
Li, Ji-Hong .
IFAC PAPERSONLINE, 2019, 52 (16) :174-179
[6]   Adaptive Neural Network Control of AUVs With Control Input Nonlinearities Using Reinforcement Learning [J].
Cui, Rongxin ;
Yang, Chenguang ;
Li, Yang ;
Sharma, Sanjay .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2017, 47 (06) :1019-1029
[7]  
Do KD, 2009, ADV IND CONTROL, P1
[8]   Reinforcement learning based model-free optimized trajectory tracking strategy design for an AUV [J].
Duan, Kairong ;
Fong, Simon ;
Chen, C. L. Philip .
NEUROCOMPUTING, 2022, 469 :289-297
[9]  
Fossen TI., 2011, HDB MARINE CRAFT HYD, DOI DOI 10.1002/9781119994138
[10]  
Gaskett C., 1999, P AUSTR C ROBOTICS A