Improving the Critic Learning for Event-Based Nonlinear H∞ Control Design

被引:81
作者
Wang, Ding [1 ,2 ,3 ]
He, Haibo [3 ]
Liu, Derong [4 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Management & Control Complex Syst, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Comp & Control Engn, Beijing 100049, Peoples R China
[3] Univ Rhode Isl, Dept Elect Comp & Biomed Engn, Kingston, RI 02881 USA
[4] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金; 美国国家科学基金会;
关键词
H-infinity control; adaptive systems; adaptive/approximate dynamic programming; critic network; event-based design; learning criterion; neural control; CONTINUOUS-TIME SYSTEMS; STATE-FEEDBACK CONTROL; TRACKING CONTROL; ALGORITHM; ITERATION;
D O I
10.1109/TCYB.2017.2653800
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we aim at improving the critic learning criterion to cope with the event-based nonlinear H-infinity state feedback control design. First of all, the H-infinity control problem is regarded as a two-player zero-sum game and the adaptive critic mechanism is used to achieve the minimax optimization under event-based environment. Then, based on an improved updating rule, the event-based optimal control law and the time-based worst-case disturbance law are obtained approximately by training a single critic neural network. The initial stabilizing control is no longer required during the implementation process of the new algorithm. Next, the closed-loop system is formulated as an impulsive model and its stability issue is handled by incorporating the improved learning criterion. The infamous Zeno behavior of the present event-based design is also avoided through theoretical analysis on the lower bound of the minimal intersample time. Finally, the applications to an aircraft dynamics and a robot arm plant are carried out to verify the efficient performance of the present novel design method.
引用
收藏
页码:3417 / 3428
页数:12
相关论文
共 38 条
[1]  
[Anonymous], 1992, HDB INTELLIGENT CONT
[2]   Decentralized Adaptive Optimal Control of Large-Scale Systems With Application to Power Systems [J].
Bian, Tao ;
Jiang, Yu ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2015, 62 (04) :2439-2447
[3]   Containment Control of Multiagent Systems With Dynamic Leaders Based on a PIn-Type Approach [J].
Cheng, Long ;
Wang, Yunpeng ;
Ren, Wei ;
Hou, Zeng-Guang ;
Tan, Min .
IEEE TRANSACTIONS ON CYBERNETICS, 2016, 46 (12) :3004-3017
[4]  
Dierks T, 2010, P AMER CONTR CONF, P1568
[5]   Optimal control of unknown affine nonlinear discrete-time systems using offline-trained neural networks with proof of convergence [J].
Dierks, Travis ;
Thumati, Balaje T. ;
Jagannathan, S. .
NEURAL NETWORKS, 2009, 22 (5-6) :851-860
[6]   Adaptive Dynamic Programming and Adaptive Optimal Output Regulation of Linear Systems [J].
Gao, Weinan ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2016, 61 (12) :4164-4169
[7]   Adaptive Neural Network Control of a Marine Vessel With Constraints Using the Asymmetric Barrier Lyapunov Function [J].
He, Wei ;
Yin, Zhao ;
Sun, Changyin .
IEEE TRANSACTIONS ON CYBERNETICS, 2017, 47 (07) :1641-1651
[8]   Finite-Horizon Control-Constrained Nonlinear Optimal Control Using Single Network Adaptive Critics [J].
Heydari, Ali ;
Balakrishnan, Sivasubramanya N. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2013, 24 (01) :145-157
[9]   Global Adaptive Dynamic Programming for Continuous-Time Nonlinear Systems [J].
Jiang, Yu ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2015, 60 (11) :2917-2929
[10]   Online adaptive approximate optimal tracking control with simplified dual approximation structure for continuous-time unknown nonlinear systems [J].
Na, Jing ;
Herrmann, Guido .
IEEE/CAA Journal of Automatica Sinica, 2014, 1 (04) :412-422