Approximate Optimal Robust Tracking Control Based on State Error and Derivative Without Initial Admissible Input

被引:8
作者
Li, Dongdong [1 ,2 ]
Dong, Jiuxiang [1 ,2 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Key Lab Vibrat & Control Aeroprop Syst, Minist Educ China, Shenyang 110819, Peoples R China
[2] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110819, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2024年 / 54卷 / 02期
基金
中国国家自然科学基金;
关键词
Adaptive dynamic programming (ADP); neural network (NN); optimal tracking control (OTC); policy iteration (PI); robust tracking control (RTC); UNKNOWN NONLINEAR-SYSTEMS; ADAPTIVE OPTIMAL-CONTROL; FEEDBACK; DESIGN;
D O I
10.1109/TSMC.2023.3320653
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this article, a robust tracking control (RTC) problem based on adaptive dynamic programming (ADP) is studied. A cost function based on the state error and derivative is used. A policy iteration (PI) algorithm that can converge is proposed to approximate the solution of the Hamilton-Jacobi- Bellman (HJB) gradually, and the feasibility of the cost function is verified. Since there is no steady-state quadratic term uTRu in the integrated part of the cost function, the boundedness of the cost function can be guaranteed without the discount factor. Unlike traditional ADP based on the discount cost function, it is theoretically proven that the error system can be asymptotically stable if the control input and cost function are iterated to optimal values. Because the error derivative is directly optimized, the excessive change of the error is suppressed. A critic neural network (NN) is used to approximate the optimal cost function. Considering the optimal gradient descent direction of the approximation error, an approximate optimal RTC algorithm without initial admissible input is derived. The difficulties of stability analysis caused by using the error derivative cost function are solved. The PI algorithm is proved to be converged and the error signal is proved to be uniformly ultimately bounded (UUB). Finally, the algorithm is verified to be effective via simulation.
引用
收藏
页码:1059 / 1069
页数:11
相关论文
共 48 条
[1]   Nearly optimal control laws for nonlinear systems with saturating actuators using a neural network HJB approach [J].
Abu-Khalaf, M ;
Lewis, FL .
AUTOMATICA, 2005, 41 (05) :779-791
[2]   Optimal Transmission Power Scheduling of Networked Control Systems Via Fuzzy Adaptive Dynamic Programming [J].
An, Liwei ;
Yang, Guang-Hong .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2021, 29 (06) :1629-1639
[3]  
Apostol T.M., 1974, Mathematical Analysis, V2nd
[4]   Data-Driven Practical Cooperative Output Regulation Under Actuator Faults and DoS Attacks [J].
Deng, Chao ;
Gao, Weinan ;
Wen, Changyun ;
Chen, Zhiyong ;
Wang, Wei .
IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (11) :7417-7428
[5]  
Dierks T, 2010, P AMER CONTR CONF, P1568
[6]   Adaptive Actor-Critic Design-Based Integral Sliding-Mode Control for Partially Unknown Nonlinear Systems With Input Disturbances [J].
Fan, Quan-Yong ;
Yang, Guang-Hong .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (01) :165-177
[7]   Learning-Based Adaptive Optimal Tracking Control of Strict-Feedback Nonlinear Systems [J].
Gao, Weinan ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (06) :2614-2624
[8]   ADP-Based Robust Resilient Control of Partially Unknown Nonlinear Systems via Cooperative Interaction Design [J].
Huang, Xin ;
Dong, Jiuxiang .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2021, 51 (12) :7466-7474
[9]   Robust Adaptive Dynamic Programming and Feedback Stabilization of Nonlinear Systems [J].
Jiang, Yu ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (05) :882-893
[10]   Actor-Critic-Based Optimal Tracking for Partially Unknown Nonlinear Discrete-Time Systems [J].
Kiumarsi, Bahare ;
Lewis, Frank L. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (01) :140-151