Robust Adaptive Dynamic Programming and Feedback Stabilization of Nonlinear Systems

被引:334
作者
Jiang, Yu [1 ]
Jiang, Zhong-Ping [1 ]
机构
[1] NYU, Dept Elect & Comp Engn, Polytech Sch Engn, Brooklyn, NY 11201 USA
基金
美国国家科学基金会;
关键词
Adaptive dynamic programming (ADP); nonlinear uncertain systems; robust optimal control; ZERO-SUM GAMES; SMALL-GAIN THEOREM; GLOBAL STABILIZATION; CONTINUOUS-TIME; PARTIAL-STATE; INPUT;
D O I
10.1109/TNNLS.2013.2294968
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper studies the robust optimal control design for a class of uncertain nonlinear systems from a perspective of robust adaptive dynamic programming (RADP). The objective is to fill up a gap in the past literature of adaptive dynamic programming (ADP) where dynamic uncertainties or unmodeled dynamics are not addressed. A key strategy is to integrate tools from modern nonlinear control theory, such as the robust redesign and the backstepping techniques as well as the nonlinear small-gain theorem, with the theory of ADP. The proposed RADP methodology can be viewed as an extension of ADP to uncertain nonlinear systems. Practical learning algorithms are developed in this paper, and have been applied to the controller design problems for a jet engine and a one-machine power system.
引用
收藏
页码:882 / 893
页数:12
相关论文
共 55 条
[1]   Neurodynamic programming and zero-sum games for constrained control systems [J].
Abu-Khalaf, Murad ;
Lewis, Frank L. ;
Huang, Jie .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (07) :1243-1252
[2]   Model-free Q-learning designs for linear discrete-time zero-sum games with application to H-infinity control [J].
Al-Tamimi, Asma ;
Lewis, Frank L. ;
Abu-Khalaf, Murad .
AUTOMATICA, 2007, 43 (03) :473-481
[3]  
[Anonymous], 1999, Nonlinear Control Systems
[4]  
[Anonymous], 2012, REINFORCEMENT LEARNI
[5]  
BAIRD LC, 1994, 1994 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOL 1-7, P2448, DOI 10.1109/ICNN.1994.374604
[6]   NEURONLIKE ADAPTIVE ELEMENTS THAT CAN SOLVE DIFFICULT LEARNING CONTROL-PROBLEMS [J].
BARTO, AG ;
SUTTON, RS ;
ANDERSON, CW .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1983, 13 (05) :834-846
[7]   Galerkin approximations of the generalized Hamilton-Jacobi-Bellman equation [J].
Beard, RW ;
Saridis, GN ;
Wen, JT .
AUTOMATICA, 1997, 33 (12) :2159-2177
[8]   DYNAMIC PROGRAMMING [J].
BELLMAN, R .
SCIENCE, 1966, 153 (3731) :34-&
[9]  
Bersekas D. P., 1996, NEURODYNAMIC PROGRAM
[10]   Reinforcement learning in continuous time and space [J].
Doya, K .
NEURAL COMPUTATION, 2000, 12 (01) :219-245