Online Model-Free Reinforcement Learning for Output Feedback Tracking Control of a Class of Discrete-Time Systems With Input Saturation

被引:2
作者
Al-Mahasneh, Ahmad Jobran [1 ]
Anavatti, Sreenatha G. [2 ]
Garratt, Matthew A. [2 ]
机构
[1] Philadelphia Univ, Mechatron Engn Dept, Amman 19392, Jordan
[2] Univ New South Wales Canberra, Sch Engn & Informat Technol, Canberra, ACT 2612, Australia
关键词
Mathematical models; Artificial neural networks; Adaptation models; System dynamics; Control systems; Optimal control; Dynamical systems; Reinforcement learning; adaptive control; nonlinear control; optimal control; NEURAL-NETWORK CONTROL; DATA-DRIVEN CONTROL; NONLINEAR-SYSTEMS; ADAPTIVE-CONTROL; FEEDFORWARD NETWORKS; DESIGN; ALGORITHM;
D O I
10.1109/ACCESS.2022.3210136
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, a new model-free Model-Actor (MA) reinforcement learning controller is developed for output feedback control of a class of discrete-time systems with input saturation constraints. The proposed controller is composed of two neural networks, namely a model-network and an actor network. The model-network is utilized to predict the output of the plant when a certain control action is applied to it. The actor network is utilized to estimate the optimal control action that is required to drive the output to the desired trajectory. The main advantages of the proposed controller over the previously proposed controllers are its ability to control systems in the absence of explicit knowledge of these systems' dynamics and its ability to start learning from scratch without any offline training. Also, it can explicitly handle the control constraints in the controller design. Comparison results with a previously published reinforcement learning output feedback controller and other controllers confirm the superiority of the proposed controller.
引用
收藏
页码:104966 / 104979
页数:14
相关论文
共 51 条
[1]  
[Anonymous], 2013, Model free adaptive control
[2]   NN Reinforcement Learning Adaptive Control for a Class of Nonstrict-Feedback Discrete-Time Systems [J].
Bai, Weiwei ;
Li, Tieshan ;
Tong, Shaocheng .
IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (11) :4573-4584
[3]   A fast training algorithm for neural networks [J].
Bilski, J ;
Rutkowski, L .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-ANALOG AND DIGITAL SIGNAL PROCESSING, 1998, 45 (06) :749-753
[4]   Model-Free Emergency Frequency Control Based on Reinforcement Learning [J].
Chen, Chunyu ;
Cui, Mingjian ;
Li, Fangxing ;
Yin, Shengfei ;
Wang, Xinan .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (04) :2336-2346
[5]   Adaptive Neural Network Control of AUVs With Control Input Nonlinearities Using Reinforcement Learning [J].
Cui, Rongxin ;
Yang, Chenguang ;
Li, Yang ;
Sharma, Sanjay .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2017, 47 (06) :1019-1029
[6]  
del Re L, 2010, LECT NOTES CONTR INF, V402, P1
[7]   Adaptive neural network control for a class of MIMO nonlinear systems with disturbances in discrete-time [J].
Ge, SS ;
Zhang, J ;
Lee, TH .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2004, 34 (04) :1630-1645
[8]   Direct and indirect reinforcement learning [J].
Guan, Yang ;
Li, Shengbo Eben ;
Duan, Jingliang ;
Li, Jie ;
Ren, Yangang ;
Sun, Qi ;
Cheng, Bo .
INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2021, 36 (08) :4439-4467
[9]   TRAINING FEEDFORWARD NETWORKS WITH THE MARQUARDT ALGORITHM [J].
HAGAN, MT ;
MENHAJ, MB .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (06) :989-993
[10]  
Haykin Simon, 2010, Neural Networks and Learning Machines, V3/E