H∞output feedback fault-tolerant control of industrial processes based on zero-sum games and off-policy Q-learning

被引:6
作者
Wang, Limin [1 ,2 ]
Jia, Linzhu [1 ]
Zhang, Ridong [3 ]
Gao, Furong [4 ]
机构
[1] Hainan Normal Univ, Sch Math & Stat, Haikou 571158, Peoples R China
[2] Guangzhou Univ, Sch Mech & Elect Engn, Guangzhou 510006, Peoples R China
[3] Hangzhou Dianzi Univ, Informat & Control Inst, Hangzhou 310018, Peoples R China
[4] Hong Kong Univ Sci & Technol, Dept Chem & Biol Engn, Hong Kong, Peoples R China
关键词
Industrial process; Fault-tolerant control; Off-policy Q-learning; Output feedback; TRACKING CONTROL; BATCH PROCESSES; TIME; DESIGN;
D O I
10.1016/j.compchemeng.2023.108421
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Traditional model-based control methods are often not applicable in industrial processes given the typical situation that model parameters are unknown, coefficient matrices are difficult to obtain, and system states are unpredictable. Accordingly, an output feedback fault-tolerant control method based on zero-sum game theory and off-policy Q-learning is presented in this study, with the aim of achieving smooth operation and good tracking performance for industrial processes that often contain sensor faults and disturbances. The specific steps are as follows. First, a system tracking error is introduced into the system to realize a novel extended model. Second, by establishing a performance index function and combining it with minimax theory, the fault-tolerant tracking control problem is converted into a zero-sum game problem. The Bellman and Riccati equations can be established after analyzing the relationship between the performance index and value functions. Then, the Qfunction is introduced, and an off-policy Q-learning algorithm is combined with the Kronecker product without knowledge of system model parameters to design an optimal controller unbiased to detection noise. Finally, the effectiveness of the algorithm is verified by considering the injection molding process as an example. The experimental results validate that the designed controller demonstrates good control and extends the range of tolerable faults while maintaining good tracking performance.
引用
收藏
页数:14
相关论文
共 35 条
  • [1] Output Feedback Q-Learning for Linear-Quadratic Discrete-Time Finite-Horizon Control Problems
    Calafiore, Giuseppe C.
    Possieri, Corrado
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (07) : 3274 - 3281
  • [2] Q-Learning: Theory and Applications
    Clifton, Jesse
    Laber, Eric
    [J]. ANNUAL REVIEW OF STATISTICS AND ITS APPLICATION, VOL 7, 2020, 2020, 7 : 279 - 301
  • [3] A Generalized Minimax Q-Learning Algorithm for Two-Player Zero-Sum Stochastic Games
    Diddigi, Raghuram Bharadwaj
    Kamanchi, Chandramouli
    Bhatnagar, Shalabh
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (09) : 4816 - 4823
  • [4] H∞ Codesign for Uncertain Nonlinear Control Systems Based on Policy Iteration Method
    Fan, Quan-Yong
    Wang, Dongsheng
    Xu, Bin
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (10) : 10101 - 10110
  • [5] Data-Driven Flotation Industrial Process Operational Optimal Control Based on Reinforcement Learning
    Jiang, Yi
    Fan, Jialu
    Chai, Tianyou
    Li, Jinna
    Lewis, Frank L.
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2018, 14 (05) : 1974 - 1989
  • [6] H∞ control of linear discrete-time systems: Off-policy reinforcement learning
    Kiumarsi, Bahare
    Lewis, Frank L.
    Jiang, Zhong-Ping
    [J]. AUTOMATICA, 2017, 78 : 144 - 152
  • [7] Off-Policy Interleaved Q-Learning: Optimal Control for Affine Nonlinear Discrete-Time Systems
    Li, Jinna
    Chai, Tianyou
    Lewis, Frank L.
    Ding, Zhengtao
    Jiang, Yi
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (05) : 1308 - 1320
  • [8] Novel MPC-Based Fault Tolerant Tracking Control Against Sensor Faults
    Li, Jun
    Zhang, Dengfeng
    Wang, Zhiquan
    [J]. ASIAN JOURNAL OF CONTROL, 2020, 22 (02) : 841 - 854
  • [9] Off-policy reinforcement learning-based novel model-free minmax fault-tolerant tracking control for industrial processes
    Li, Xueyu
    Luo, Qiuwen
    Wang, Limin
    Zhang, Ridong
    Gao, Furong
    [J]. JOURNAL OF PROCESS CONTROL, 2022, 115 : 145 - 156
  • [10] Output feedback Q-learning for discrete-time finite-horizon zero-sum games with application to the H? control
    Liu, Mingxiang
    Cai, Qianqian
    Li, Dandan
    Meng, Wei
    Fu, Minyue
    [J]. NEUROCOMPUTING, 2023, 529 : 48 - 55