Game Theoretical Reinforcement Learning for Robust H∞ Tracking Control of Discrete-Time Linear Systems with Unknown Dynamics

被引:0
作者
Wu, Hao [1 ]
Li, Shaobao [1 ]
Durdevic, Petar [2 ]
Yang, Zhenyu [2 ]
机构
[1] Yanshan Univ, Sch Elect Engn, Qinhuangdao 066004, Hebei, Peoples R China
[2] Aalborg Univ, Dept Energy Technol, DK-6700 Aalborg, Denmark
来源
2021 4TH INTERNATIONAL CONFERENCE ON INTELLIGENT AUTONOMOUS SYSTEMS (ICOIAS 2021) | 2021年
关键词
H-infinity control; game theory; reinforcement learning; zero-sum game; de-oiling hydrocyclone system;
D O I
10.1109/ICoIAS53694.2021.00058
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robust H-infinity control has been widely studied to improve control performance for industrial process control systems against disturbances. However, most of existing robust H-infinity control are model-based, and their deployment in some industrial facilities may greatly increase the installation and maintenance costs due to requiring system identification. Towards this end, a model-free robust H-infinity tracking control scheme is developed based on game theoretical reinforcement learning (RL) for discrete-time linear systems with unknown dynamics. The normal robust H-infinity tracking control problem is first modeled as a two-player zero-sum game with the controller and disturbance as the two players. A model-based solution by solving game discrete-time differential Riccati equation (GDARE) is introduced to show the solvability of the robust H-infinity tracking control problem, and then a novel off-policy RL algorithm is developed to replace the GDARE method for model-free robust H-infinity tracking control of the discrete-time linear systems with unknown dynamics. Stability of the learning algorithm is analyzed. Finally, a simulation study upon a de-oiling hydrocyclone system is conducted to demonstrate the effectiveness of the proposed algorithm.
引用
收藏
页码:290 / 295
页数:6
相关论文
共 50 条
[41]   H∞ control of switched linear discrete-time systems with polytopic uncertainties [J].
Zhang, Lixian ;
Shi, Peng ;
Boukas, El-Kebir ;
Wang, Changhong .
OPTIMAL CONTROL APPLICATIONS & METHODS, 2006, 27 (05) :273-291
[42]   Optimal Tracking Control of Linear Discrete-Time Systems Under Cyber Attacks [J].
Liu, Hao ;
Qiu, Hui .
IFAC PAPERSONLINE, 2020, 53 (02) :3545-3550
[43]   Linear Quadratic Tracking Control of Partially-Unknown Continuous-Time Systems Using Reinforcement Learning [J].
Modares, Hamidreza ;
Lewis, Frank L. .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2014, 59 (11) :3051-3056
[44]   Optimal tracking control for non-zero-sum games of linear discrete-time systems via off-policy reinforcement learning [J].
Wen, Yinlei ;
Zhang, Huaguang ;
Su, Hanguang ;
Ren, He .
OPTIMAL CONTROL APPLICATIONS & METHODS, 2020, 41 (04) :1233-1250
[45]   Reinforcement Learning Consensus Control for Discrete-Time Multi-Agent Systems [J].
Zhu, Xiaoxia ;
Yuan, Xin ;
Wang, Yuanda ;
Sun, Changyin .
PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, :6178-6182
[46]   Data-driven disturbance compensation control for discrete-time systems based on reinforcement learning [J].
Li, Lanyue ;
Li, Jinna ;
Cao, Jiangtao .
INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2025, 39 (07) :1364-1379
[47]   Optimal Output Regulation of Partially Linear Discrete-time Systems Using Reinforcement Learning [J].
Pang W.-Y. ;
Fan J.-L. ;
Jiang Y. ;
Lewis F.L. .
Zidonghua Xuebao/Acta Automatica Sinica, 2022, 48 (09) :2242-2253
[48]   Robust H infinity control for linear discrete-time systems with norm-bounded time-varying uncertainty [J].
Yuan, LS ;
Achenie, LEK ;
Jiang, WS .
SYSTEMS & CONTROL LETTERS, 1996, 27 (04) :199-208
[49]   Reinforcement Learning for Input Constrained Sub-optimal Tracking Control in Discrete-time Two-time-scale Systems [J].
Que, Xuejie ;
Wang, Zhenlei ;
Wang, Xin .
INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2023, 21 (09) :3068-3079
[50]   Reinforcement Learning for Input Constrained Sub-optimal Tracking Control in Discrete-time Two-time-scale Systems [J].
Xuejie Que ;
Zhenlei Wang ;
Xin Wang .
International Journal of Control, Automation and Systems, 2023, 21 (9) :3068-3079