Online Model-Free Reinforcement Learning for Output Feedback Tracking Control of a Class of Discrete-Time Systems With Input Saturation

被引:2
作者
Al-Mahasneh, Ahmad Jobran [1 ]
Anavatti, Sreenatha G. [2 ]
Garratt, Matthew A. [2 ]
机构
[1] Philadelphia Univ, Mechatron Engn Dept, Amman 19392, Jordan
[2] Univ New South Wales Canberra, Sch Engn & Informat Technol, Canberra, ACT 2612, Australia
关键词
Mathematical models; Artificial neural networks; Adaptation models; System dynamics; Control systems; Optimal control; Dynamical systems; Reinforcement learning; adaptive control; nonlinear control; optimal control; NEURAL-NETWORK CONTROL; DATA-DRIVEN CONTROL; NONLINEAR-SYSTEMS; ADAPTIVE-CONTROL; FEEDFORWARD NETWORKS; DESIGN; ALGORITHM;
D O I
10.1109/ACCESS.2022.3210136
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, a new model-free Model-Actor (MA) reinforcement learning controller is developed for output feedback control of a class of discrete-time systems with input saturation constraints. The proposed controller is composed of two neural networks, namely a model-network and an actor network. The model-network is utilized to predict the output of the plant when a certain control action is applied to it. The actor network is utilized to estimate the optimal control action that is required to drive the output to the desired trajectory. The main advantages of the proposed controller over the previously proposed controllers are its ability to control systems in the absence of explicit knowledge of these systems' dynamics and its ability to start learning from scratch without any offline training. Also, it can explicitly handle the control constraints in the controller design. Comparison results with a previously published reinforcement learning output feedback controller and other controllers confirm the superiority of the proposed controller.
引用
收藏
页码:104966 / 104979
页数:14
相关论文
共 51 条
[11]   Reinforcement learning-based output feedback control of nonlinear systems with input constraints [J].
He, P ;
Jagannathan, S .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2005, 35 (01) :150-154
[12]   Adaptive Neural Network Control of a Flapping Wing Micro Aerial Vehicle With Disturbance Observer [J].
He, Wei ;
Yan, Zichen ;
Sun, Changyin ;
Chen, Yunan .
IEEE TRANSACTIONS ON CYBERNETICS, 2017, 47 (10) :3452-3465
[13]   Adaptive Neural Impedance Control of a Robotic Manipulator With Input Saturation [J].
He, Wei ;
Dong, Yiting ;
Sun, Changyin .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2016, 46 (03) :334-344
[14]   Adaptive Neural Network Control of an Uncertain Robot With Full-State Constraints [J].
He, Wei ;
Chen, Yuhao ;
Yin, Zhao .
IEEE TRANSACTIONS ON CYBERNETICS, 2016, 46 (03) :620-629
[15]   MULTILAYER FEEDFORWARD NETWORKS ARE UNIVERSAL APPROXIMATORS [J].
HORNIK, K ;
STINCHCOMBE, M ;
WHITE, H .
NEURAL NETWORKS, 1989, 2 (05) :359-366
[16]   From model-based control to data-driven control: Survey, classification and perspective [J].
Hou, Zhong-Sheng ;
Wang, Zhuo .
INFORMATION SCIENCES, 2013, 235 :3-35
[17]   Data-Driven Control and Learning Systems [J].
Hou, Zhongsheng ;
Gao, Huijun ;
Lewis, Frank L. .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2017, 64 (05) :4070-4075
[18]   Plume Tracing via Model-Free Reinforcement Learning Method [J].
Hu, Hangkai ;
Song, Shiji ;
Chen, C. L. Phillip .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (08) :2515-2527
[19]   Neural-Network-Based Robust Control Schemes for Nonlinear Multiplayer Systems With Uncertainties via Adaptive Dynamic Programming [J].
Jiang, He ;
Zhang, Huaguang ;
Luo, Yanhong ;
Han, Ji .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2019, 49 (03) :579-588
[20]  
Kamalapurkar R, 2018, COMMUN CONTROL ENG, P1, DOI 10.1007/978-3-319-78384-0_1