Reinforcement learning neural-network-based controller for nonlinear discrete-time systems with input constraints

被引:189
作者
He, Pingan [1 ]
Jagannathan, S. [1 ]
机构
[1] Univ Missouri, Dept Elect & Comp Engn, Rolla, MO 65409 USA
来源
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS | 2007年 / 37卷 / 02期
基金
美国国家科学基金会;
关键词
approximate dynamic programming; neural network control; optimal control; reinforcement learning;
D O I
10.1109/TSMCB.2006.883869
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A novel adaptive-critic-based neural network (NN) controller in discrete time is designed to deliver a desired tracking performance for a class of nonlinear systems in the presence of actuator constraints. The constraints of the actuator are treated in the controller design as the saturation nonlinearity. The adaptive critic NN controller architecture based on state feedback includes two NNs: the critic NN is used to approximate the "strategic" utility function, whereas the action NN is employed to minimize both the strategic utility function and the unknown nonlinear dynamic estimation errors. The critic and action NN weight updates are derived by minimizing certain quadratic performance indexes. Using the Lyapunov approach and with novel weight updates, the uniformly ultimate boundedness of the closed-loop tracking error and weight estimates is shown in the presence of NN approximation errors and bounded unknown disturbances. The proposed NN controller works in the presence of multiple nonlinearities, unlike other schemes that normally approximate one nonlinearity. Moreover, the adaptive critic NN controller does not require an explicit offline training phase, and the NN weights can be initialized at zero or random. Simulation results justify the theoretical analysis.
引用
收藏
页码:425 / 436
页数:12
相关论文
共 25 条