Learning-based adaptive optimal output regulation of linear and nonlinear systems: an overview

被引:17
作者
Gao, Weinan [1 ]
Jiang, Zhong-Ping [2 ]
机构
[1] Florida Inst Technol, Coll Engn & Sci, Dept Mech & Civil Engn, 150 W Univ Blvd, Melbourne, FL 32901 USA
[2] NYU, Tandon Sch Engn, Dept Elect & Comp Engn, 6 MetroTech Ctr, Brooklyn, NY 11201 USA
基金
美国国家科学基金会;
关键词
Adaptive optimal output regulation; Adaptive dynamic programming; Reinforcement learning; Learning-based control; INTERNAL-MODEL PRINCIPLE; MULTIAGENT SYSTEMS; NEURAL-NETWORKS; VALUE-ITERATION; CRUISE CONTROL; STABILITY; CONTROLLERS; TRACKING; GAME; GO;
D O I
10.1007/s11768-022-00081-3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper reviews recent developments in learning-based adaptive optimal output regulation that aims to solve the problem of adaptive and optimal asymptotic tracking with disturbance rejection. The proposed framework aims to bring together two separate topics-output regulation and adaptive dynamic programming-that have been under extensive investigation due to their broad applications in modern control engineering. Under this framework, one can solve optimal output regulation problems of linear, partially linear, nonlinear, and multi-agent systems in a data-driven manner. We will also review some practical applications based on this framework, such as semi-autonomous vehicles, connected and autonomous vehicles, and nonlinear oscillators.
引用
收藏
页码:1 / 19
页数:19
相关论文
共 95 条
[61]   Output regulation for linear systems via adaptive. internal model [J].
Marino, R ;
Tomei, P .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2003, 48 (12) :2199-2202
[62]   Adaptive output regulation for minimum-phase systems with unknown relative degree [J].
Marino, Riccardo ;
Tomei, Patrizio .
AUTOMATICA, 2021, 130
[63]   Adaptive dynamic programming [J].
Murray, JJ ;
Cox, CJ ;
Lendaris, GG ;
Saeks, R .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2002, 32 (02) :140-153
[64]   Distributed predictive cruise control based on reinforcement learning and validation on microscopic traffic simulation [J].
Mynuddin, Mohammed ;
Gao, Weinan .
IET INTELLIGENT TRANSPORT SYSTEMS, 2020, 14 (05) :270-277
[65]   Nonlinear Constrained optimal Control of Wave Energy Converters With Adaptive Dynamic Programming [J].
Na, Jing ;
Wang, Bin ;
Li, Guang ;
Zhan, Siyuan ;
He, Wei .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2019, 66 (10) :7904-7915
[66]   Reinforcement learning and non-zero-sum game output regulation for multi-player linear uncertain systems [J].
Odekunle, Adedapo ;
Gao, Weinan ;
Davari, Masoud ;
Jiang, Zhong-Ping .
AUTOMATICA, 2020, 112
[67]  
Odekunle A, 2019, IEEE-CAA J AUTOMATIC, V6, P1108, DOI [10.1109/JAS.2019.1911678, 10.1109/JAS.2019.19117678]
[68]   Robust Policy Iteration for Continuous-Time Linear Quadratic Regulation [J].
Pang, Bo ;
Bian, Tao ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (01) :504-511
[69]  
Powell M. J. D., 1981, APPROXIMATION THEORY, DOI DOI 10.1016/j.biosystems.2005.06.016
[70]   Numerical method for the solution of the regulator equation with application to nonlinear tracking [J].
Rehak, Branislav ;
Celikovsky, Sergej .
AUTOMATICA, 2008, 44 (05) :1358-1365