Reinforcement Q-learning algorithm for H∞ tracking control of discrete-time Markov jump systems

被引:0
|
作者
Shi, Jiahui [1 ]
He, Dakuo [1 ,2 ]
Zhang, Qiang [1 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Shenyang, Peoples R China
[2] State Key Lab Synthet Automat Proc Ind, Shenyang, Peoples R China
基金
中国国家自然科学基金;
关键词
Markov jump systems; reinforcement learning; H-infinity tracking control; tracking game algebraicRiccati equation; NONLINEAR-SYSTEMS; DESIGN;
D O I
10.1080/00207721.2024.2395928
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, the H-infinity tracking control problem of linear discrete-time Markov jump systems is studied by using the data-based reinforcement learning method. Specifically, a new performance index function is established by using Markov chain and weighted sum technique, and thus the tracking game algebraic Riccati equation with weight vector and discount factor is obtained. A Q-learning algorithm is proposed to solve the tracking game algebra Riccati equation problem online without knowing the information of the system model. In addition, the convergence analysis of the algorithm is given, and it is proved that the added probing noise will not bias the algorithm. Finally, two simulation examples are given to verify the effectiveness of the proposed algorithm.
引用
收藏
页码:502 / 523
页数:22
相关论文
共 50 条
  • [1] Reinforcement Q-Learning Algorithm for H∞ Tracking Control of Unknown Discrete-Time Linear Systems
    Peng, Yunjian
    Chen, Qian
    Sun, Weijie
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2020, 50 (11): : 4109 - 4122
  • [2] H∞ Tracking learning control for discrete-time Markov jump systems: A parallel off-policy reinforcement learning
    Zhang, Xuewen
    Xia, Jianwei
    Wang, Jing
    Chen, Xiangyong
    Shen, Hao
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2023, 360 (18): : 14878 - 14890
  • [3] Fuzzy H8 Control of Discrete-Time Nonlinear Markov Jump Systems via a Novel Hybrid Reinforcement Q-Learning Method
    Wang, Jing
    Wu, Jiacheng
    Shen, Hao
    Cao, Jinde
    Rutkowski, Leszek
    IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (11) : 7380 - 7391
  • [4] Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics
    Kiumarsi, Bahare
    Lewis, Frank L.
    Modares, Hamidreza
    Karimpour, Ali
    Naghibi-Sistani, Mohammad-Bagher
    AUTOMATICA, 2014, 50 (04) : 1167 - 1175
  • [5] H∞ Tracking Control for Linear Discrete-Time Systems: Model-Free Q-Learning Designs
    Yang, Yunjie
    Wan, Yan
    Zhu, Jihong
    Lewis, Frank L.
    IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (01): : 175 - 180
  • [6] Minimax Q-learning design for H∞ control of linear discrete-time systems
    Li, Xinxing
    Xi, Lele
    Zha, Wenzhong
    Peng, Zhihong
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2022, 23 (03) : 438 - 451
  • [7] Stochastic linear quadratic optimal tracking control for discrete-time systems with delays based on Q-learning algorithm
    Tan, Xufeng
    Li, Yuan
    Liu, Yang
    AIMS MATHEMATICS, 2023, 8 (05): : 10249 - 10265
  • [8] H∞ tracking control for linear discrete-time systems via reinforcement learning
    Liu, Ying-Ying
    Wang, Zhan-Shan
    Shi, Zhan
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2020, 30 (01) : 282 - 301
  • [9] Q-learning solution for optimal consensus control of discrete-time multiagent systems using reinforcement learning
    Mu, Chaoxu
    Zhao, Qian
    Gao, Zhongke
    Sun, Changyin
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2019, 356 (13): : 6946 - 6967
  • [10] Off-policy reinforcement learning for tracking control of discrete-time Markov jump linear systems with completely unknown dynamics
    Huang Z.
    Tu Y.
    Fang H.
    Wang H.
    Zhang L.
    Shi K.
    He S.
    Journal of the Franklin Institute, 2023, 360 (03) : 2361 - 2378