Adaptive reinforcement learning fault-tolerant control for AUVs With thruster faults based on the integral extended state observer

被引:21
作者
Li, Zhifu [1 ,2 ]
Wang, Ming [1 ]
Ma, Ge [1 ,2 ]
Zou, Tao [1 ,2 ]
机构
[1] Guangzhou Univ, Sch Mech & Elect Engn, Guangzhou 510006, Peoples R China
[2] Guangzhou Univ, Guangzhou Key Lab Mfg Proc Automat, Guangzhou 510006, Peoples R China
基金
中国国家自然科学基金;
关键词
Integral extended state observer (IESO); Autonomous underwater vehicles (AUVs); Fault-tolerant control (FTC); Thruster faults; Reinforcement learning (RL); DISCRETE-TIME-SYSTEMS; TRACKING CONTROL; NONLINEAR-SYSTEMS; VEHICLES;
D O I
10.1016/j.oceaneng.2023.113722
中图分类号
U6 [水路运输]; P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
In this paper, a reinforcement learning (RL) fault-tolerant control (FTC) method is proposed for trajectory tracking of autonomous underwater vehicles (AUVs) with thruster faults. To deal with the thruster fault, unknown disturbance and model uncertainty, a new integral extended state observer (IESO) for fault diagnosis observation is proposed, which uses a conventional ESO to estimate the total system uncertainty, and introduces an integral mechanism to mitigate the effect of estimation error further. Thus, the problem that the estimation error caused by the traditional ESO leads to the decline of the fault-tolerant capability of the FTC system is solved. Then, to solve the problem of integral saturation due to the introduction of the integral term, the integral term is limited after the thruster fault of the AUV. Furthermore, based on the actor-critic structure of RL, a PD-like feedback controller is designed to realize the FTC of AUV in the face of thruster fault by using the total uncertainty of the IESO scheme. And the input saturation of the thruster is considered, and an auxiliary variable system is used to handle the control truncation between saturated and unsaturated inputs. Based on the Lyapunov method, the stability of the closed-loop system is analyzed and proved. Finally, the proposed method is verified to have good fault tolerance and robustness by simulation and underwater experiments.
引用
收藏
页数:12
相关论文
共 42 条
[1]  
[Anonymous], 1995, Control Decis., DOI DOI 10.13195/J.CD.1995.01.85.HANJQ.020
[2]   NN Reinforcement Learning Adaptive Control for a Class of Nonstrict-Feedback Discrete-Time Systems [J].
Bai, Weiwei ;
Li, Tieshan ;
Tong, Shaocheng .
IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (11) :4573-4584
[3]  
Cao Wei, 2013, Control and Decision, V28, P137
[4]   Single critic network based fault-tolerant tracking control for underactuated AUV with actuator fault [J].
Che, Gaofeng .
OCEAN ENGINEERING, 2022, 254
[5]   Rapid Sensor Fault Diagnosis for a Class of Nonlinear Systems via Deterministic Learning [J].
Chen, Tianrui ;
Zhu, Zejian ;
Wang, Cong ;
Dong, ZhaoYang .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (12) :7743-7754
[6]   Adaptive trajectory tracking control for remotely operated vehicles considering thruster dynamics and saturation constraints [J].
Chu, Zhenzhong ;
Xiang, Xianbo ;
Zhu, Daqi ;
Luo, Chaomin ;
Xie, De .
ISA TRANSACTIONS, 2020, 100 :28-37
[7]   Adaptive Neural Network Control of AUVs With Control Input Nonlinearities Using Reinforcement Learning [J].
Cui, Rongxin ;
Yang, Chenguang ;
Li, Yang ;
Sharma, Sanjay .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2017, 47 (06) :1019-1029
[8]   Active fault-tolerant control scheme of aerial manipulators with actuator faults [J].
Ding, Ya-dong ;
Wang, Yao-yao ;
Jiang, Su-rong ;
Chen, Bai .
JOURNAL OF CENTRAL SOUTH UNIVERSITY, 2021, 28 (03) :771-783
[9]   AUV position tracking and trajectory control based on fast-deployed deep reinforcement learning method [J].
Fang, Yuan ;
Huang, Zhenwei ;
Pu, Jinyun ;
Zhang, Jinsong .
OCEAN ENGINEERING, 2022, 245
[10]   Adaptive neural control of uncertain MIMO nonlinear systems [J].
Ge, SS ;
Wang, C .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2004, 15 (03) :674-692