Game Theoretical Reinforcement Learning for Robust H∞ Tracking Control of Discrete-Time Linear Systems with Unknown Dynamics

被引:0
作者
Wu, Hao [1 ]
Li, Shaobao [1 ]
Durdevic, Petar [2 ]
Yang, Zhenyu [2 ]
机构
[1] Yanshan Univ, Sch Elect Engn, Qinhuangdao 066004, Hebei, Peoples R China
[2] Aalborg Univ, Dept Energy Technol, DK-6700 Aalborg, Denmark
来源
2021 4TH INTERNATIONAL CONFERENCE ON INTELLIGENT AUTONOMOUS SYSTEMS (ICOIAS 2021) | 2021年
关键词
H-infinity control; game theory; reinforcement learning; zero-sum game; de-oiling hydrocyclone system;
D O I
10.1109/ICoIAS53694.2021.00058
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robust H-infinity control has been widely studied to improve control performance for industrial process control systems against disturbances. However, most of existing robust H-infinity control are model-based, and their deployment in some industrial facilities may greatly increase the installation and maintenance costs due to requiring system identification. Towards this end, a model-free robust H-infinity tracking control scheme is developed based on game theoretical reinforcement learning (RL) for discrete-time linear systems with unknown dynamics. The normal robust H-infinity tracking control problem is first modeled as a two-player zero-sum game with the controller and disturbance as the two players. A model-based solution by solving game discrete-time differential Riccati equation (GDARE) is introduced to show the solvability of the robust H-infinity tracking control problem, and then a novel off-policy RL algorithm is developed to replace the GDARE method for model-free robust H-infinity tracking control of the discrete-time linear systems with unknown dynamics. Stability of the learning algorithm is analyzed. Finally, a simulation study upon a de-oiling hydrocyclone system is conducted to demonstrate the effectiveness of the proposed algorithm.
引用
收藏
页码:290 / 295
页数:6
相关论文
共 50 条
[31]   Reinforcement Learning Control for Discrete-time Systems over Fading Channels [J].
Lai, Jing ;
Xiong, Junlin .
IFAC PAPERSONLINE, 2023, 56 (02) :4834-4839
[32]   Model-free H∞ control design for unknown linear discrete-time systems via Q-learning with LMI [J].
Kim, J. -H. ;
Lewis, F. L. .
AUTOMATICA, 2010, 46 (08) :1320-1326
[33]   Optimal tracking control for completely unknown nonlinear discrete-time Markov jump systems using data-based reinforcement learning method [J].
Jiang, He ;
Zhang, Huaguang ;
Luo, Yanhong ;
Wang, Junyi .
NEUROCOMPUTING, 2016, 194 :176-182
[34]   Robust discrete-time H∞,-optimal tracking with preview [J].
Cohen, A ;
Shaked, U .
INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 1998, 8 (01) :29-37
[35]   Generalized Policy Iteration-based Reinforcement Learning Algorithm for Optimal Control of Unknown Discrete-time Systems [J].
Lin, Mingduo ;
Zhao, Bo ;
Liu, Derong ;
Liu, Xi ;
Luo, Fangchao .
PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, :3650-3655
[36]   Online Adaptive Policy Learning Algorithm for H∞ State Feedback Control of Unknown Affine Nonlinear Discrete-Time Systems [J].
Zhang, Huaguang ;
Qin, Chunbin ;
Jiang, Bin ;
Luo, Yanhong .
IEEE TRANSACTIONS ON CYBERNETICS, 2014, 44 (12) :2706-2718
[37]   Robust Control for Discrete-Time Networked Control Systems [J].
Wu, Dongxiao ;
Wu, Jun ;
Chen, Sheng .
2010 8TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2010, :3532-3537
[38]   Based on off-policy Q-learning: Optimal tracking control for unknown linear discrete-time systems under deception attacks [J].
Song, Xing-Xing ;
Chu, Zhao-Bi .
Kongzhi yu Juece/Control and Decision, 2025, 40 (05) :1641-1650
[39]   Distributed Optimal Tracking Control of Discrete-Time Multiagent Systems via Event-Triggered Reinforcement Learning [J].
Peng, Zhinan ;
Luo, Rui ;
Hu, Jiangping ;
Shi, Kaibo ;
Ghosh, Bijoy Kumar .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (09) :3689-3700
[40]   Adaptive Fault-Tolerant Tracking Control for Discrete-Time Multiagent Systems via Reinforcement Learning Algorithm [J].
Li, Hongyi ;
Wu, Ying ;
Chen, Mou .
IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (03) :1163-1174