Robust Optimal Control Scheme for Unknown Constrained-Input Nonlinear Systems via a Plug-n-Play Event-Sampled Critic-Only Algorithm

被引:113
作者
Zhang, Huaguang [1 ,2 ]
Zhang, Kun [2 ]
Xiao, Geyang [2 ]
Jiang, He [2 ]
机构
[1] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110004, Peoples R China
[2] Northeastern Univ, Sch Informat Sci & Engn, Shenyang 110004, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2020年 / 50卷 / 09期
基金
中国国家自然科学基金; 国家高技术研究发展计划(863计划);
关键词
Optimal control; Artificial neural networks; Heuristic algorithms; Approximation algorithms; Robust control; Nonlinear dynamical systems; Adaptive dynamic programming (ADP); continuous-time nonlinear system; input constraint; neural network (NN); optimal control; robust control; OPTIMAL TRACKING CONTROL; INFINITY STATE-FEEDBACK; ZERO-SUM GAMES; POLICY ITERATION; LEARNING ALGORITHM; ARCHITECTURE; DESIGN;
D O I
10.1109/TSMC.2018.2889377
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, a novel event-sampled robust optimal controller is proposed for a class of continuous-time constrained-input nonlinear systems with unknown dynamics. In order to solve the robust optimal control problem, an online data-driven identifier is established to construct the system dynamics, and an event-sampled critic-only adaptive dynamic programming method is developed to replace the conventional time-driven actor-critic structure. The designed online identification method runs during the solving process and is not applied as a priori part for the solutions, which simplifies the architecture and reduces computational load. The proposed robust optimal control algorithm tunes the parameters of critic-only neural network (NN) by event-triggering condition and runs in a plug-n-play framework without system functions, where fewer transmissions and less computation are required as all the measurements received simultaneously. Based on the novel design, the stability of system and the convergence of critic NN are demonstrated by Lyapunov theory, where the state is asymptotically stable and weight error is guaranteed to be uniformly ultimately bounded. Finally, the applications in a basic nonlinear system and the complex rotational-translational actuator problem demonstrate the effectiveness of the proposed method.
引用
收藏
页码:3169 / 3180
页数:12
相关论文
共 47 条
[1]   Policy iterations on the Hamilton-Jacobi-Isaacs equation for H∞ state feedback control with input saturation [J].
Abu-Khalaf, Murad ;
Lewis, Frank L. ;
Huang, Jie .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2006, 51 (12) :1989-1995
[2]   Bounded robust control of nonlinear systems using neural network-based HJB solution [J].
Adhyaru, Dipak M. ;
Kar, I. N. ;
Gopal, M. .
NEURAL COMPUTING & APPLICATIONS, 2011, 20 (01) :91-103
[3]  
[Anonymous], 2016, LECT NOTES COMPUT SC, DOI DOI 10.1007/978-3-319-40663-3_31
[4]   A novel actor-critic-identifier architecture for approximate optimal control of uncertain nonlinear systems [J].
Bhasin, S. ;
Kamalapurkar, R. ;
Johnson, M. ;
Vamvoudakis, K. G. ;
Lewis, F. L. ;
Dixon, W. E. .
AUTOMATICA, 2013, 49 (01) :82-92
[5]  
Ding J., 2010, P INT J C NEUR NETW, P1
[6]   Adaptive Event-Triggered Control Based on Heuristic Dynamic Programming for Nonlinear Discrete-Time Systems [J].
Dong, Lu ;
Zhong, Xiangnan ;
Sun, Changyin ;
He, Haibo .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (07) :1594-1605
[7]   Single network adaptive critic design for power system stabilisers [J].
Gurrala, G. ;
Sen, I. ;
Padhi, R. .
IET GENERATION TRANSMISSION & DISTRIBUTION, 2009, 3 (09) :850-858
[8]   Finite-Horizon Control-Constrained Nonlinear Optimal Control Using Single Network Adaptive Critics [J].
Heydari, Ali ;
Balakrishnan, Sivasubramanya N. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2013, 24 (01) :145-157
[9]   Global Adaptive Dynamic Programming for Continuous-Time Nonlinear Systems [J].
Jiang, Yu ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2015, 60 (11) :2917-2929
[10]   Optimal Codesign of Nonlinear Control Systems Based on a Modified Policy Iteration Method [J].
Jiang, Yu ;
Wang, Yebin ;
Bortoff, Scott A. ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (02) :409-414