Game Theoretical Reinforcement Learning for Robust H∞ Tracking Control of Discrete-Time Linear Systems with Unknown Dynamics

被引:0
作者
Wu, Hao [1 ]
Li, Shaobao [1 ]
Durdevic, Petar [2 ]
Yang, Zhenyu [2 ]
机构
[1] Yanshan Univ, Sch Elect Engn, Qinhuangdao 066004, Hebei, Peoples R China
[2] Aalborg Univ, Dept Energy Technol, DK-6700 Aalborg, Denmark
来源
2021 4TH INTERNATIONAL CONFERENCE ON INTELLIGENT AUTONOMOUS SYSTEMS (ICOIAS 2021) | 2021年
关键词
H-infinity control; game theory; reinforcement learning; zero-sum game; de-oiling hydrocyclone system;
D O I
10.1109/ICoIAS53694.2021.00058
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robust H-infinity control has been widely studied to improve control performance for industrial process control systems against disturbances. However, most of existing robust H-infinity control are model-based, and their deployment in some industrial facilities may greatly increase the installation and maintenance costs due to requiring system identification. Towards this end, a model-free robust H-infinity tracking control scheme is developed based on game theoretical reinforcement learning (RL) for discrete-time linear systems with unknown dynamics. The normal robust H-infinity tracking control problem is first modeled as a two-player zero-sum game with the controller and disturbance as the two players. A model-based solution by solving game discrete-time differential Riccati equation (GDARE) is introduced to show the solvability of the robust H-infinity tracking control problem, and then a novel off-policy RL algorithm is developed to replace the GDARE method for model-free robust H-infinity tracking control of the discrete-time linear systems with unknown dynamics. Stability of the learning algorithm is analyzed. Finally, a simulation study upon a de-oiling hydrocyclone system is conducted to demonstrate the effectiveness of the proposed algorithm.
引用
收藏
页码:290 / 295
页数:6
相关论文
共 50 条
[21]   Learning Optimal Control Policy for Unknown Discrete-Time Systems [J].
Lai, Jing ;
Xiong, Junlin .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (11) :4191-4195
[22]   H∞ Robust Controller for Discrete-Time Linear Systems Under Control and Output Constraints [J].
Nogueira, Alvaro ;
Araujo, Humberto X. ;
Oliveira, Gustavo H. C. .
2009 IEEE INTERNATIONAL CONFERENCE ON CONTROL AND AUTOMATION, VOLS 1-3, 2009, :495-+
[23]   Compositional Reinforcement Learning for Discrete-Time Stochastic Control Systems [J].
Lavaei, Abolfazl ;
Perez, Mateo ;
Kazemi, Milad ;
Somenzi, Fabio ;
Soudjani, Sadegh ;
Trivedi, Ashutosh ;
Zamani, Majid .
IEEE OPEN JOURNAL OF CONTROL SYSTEMS, 2023, 2 :425-438
[24]   DTSRL: Efficient reinforcement learning for approximate optimal tracking control of discrete-time nonlinear systems [J].
Fu, Hao ;
Zhou, Shuai ;
Liu, Wei .
COMMUNICATIONS IN NONLINEAR SCIENCE AND NUMERICAL SIMULATION, 2025, 150
[25]   Reinforcement Learning-Based Tracking Control for a Class of Discrete-Time Systems With Actuator Fault [J].
Liu, Yingying ;
Wang, Zhanshan .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (06) :2827-2831
[26]   Reinforcement Learning for H∞ Optimal Control of Unknown Continuous-Time Linear Systems [J].
Li, Hongyang ;
Wei, Qinglai ;
Tan, Xiangmin .
IEEE TRANSACTIONS ON CYBERNETICS, 2025, 55 (05) :2379-2389
[27]   H∞ control for discrete-time Markovian jump linear systems with partly unknown transition probabilities [J].
Zhang, Lixian ;
Boukas, El-Kebir .
INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2009, 19 (08) :868-883
[28]   Robust H∞ control for linear discrete-time systems with norm-bounded nonlinear uncertainties [J].
Shi, P ;
Shue, SP .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1999, 44 (01) :108-111
[29]   Robust value iteration for optimal control of discrete-time linear systems [J].
Lai, Jing ;
Xiong, Junlin .
AUTOMATICA, 2025, 174
[30]   Online optimal and adaptive integral tracking control for varying discrete-time systems using reinforcement learning [J].
Sanusi, Ibrahim ;
Mills, Andrew ;
Dodd, Tony ;
Konstantopoulos, George .
INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2020, 34 (08) :971-991