Reinforcement Q-learning algorithm for H∞ tracking control of discrete-time Markov jump systems

被引:0
|
作者
Shi, Jiahui [1 ]
He, Dakuo [1 ,2 ]
Zhang, Qiang [1 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Shenyang, Peoples R China
[2] State Key Lab Synthet Automat Proc Ind, Shenyang, Peoples R China
基金
中国国家自然科学基金;
关键词
Markov jump systems; reinforcement learning; H-infinity tracking control; tracking game algebraicRiccati equation; NONLINEAR-SYSTEMS; DESIGN;
D O I
10.1080/00207721.2024.2395928
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, the H-infinity tracking control problem of linear discrete-time Markov jump systems is studied by using the data-based reinforcement learning method. Specifically, a new performance index function is established by using Markov chain and weighted sum technique, and thus the tracking game algebraic Riccati equation with weight vector and discount factor is obtained. A Q-learning algorithm is proposed to solve the tracking game algebra Riccati equation problem online without knowing the information of the system model. In addition, the convergence analysis of the algorithm is given, and it is proved that the added probing noise will not bias the algorithm. Finally, two simulation examples are given to verify the effectiveness of the proposed algorithm.
引用
收藏
页码:502 / 523
页数:22
相关论文
共 50 条
  • [21] H? Asynchronous Control for Discrete-Time Semi-Markov Jump Systems
    Jin, Liqiang
    Yin, Yanyan
    Liu, Yanqing
    Teo, Kok Lay
    Liu, Fei
    IFAC PAPERSONLINE, 2022, 55 (03): : 143 - 148
  • [22] H∞ control of discrete-time Markov jump systems with bounded transition probabilities
    Boukas, E. K.
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2009, 30 (05): : 477 - 494
  • [23] Optimal tracking control for discrete-time modal persistent dwell time switched systems based on Q-learning
    Zhang, Xuewen
    Wang, Yun
    Xia, Jianwei
    Li, Feng
    Shen, Hao
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2023, 44 (06): : 3327 - 3341
  • [24] Tracking control of discrete-time Markovian jump systems
    Tian, Guangtai
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2020, 51 (15) : 3070 - 3080
  • [25] H∞ Tracking Control of Unknown Discrete-Time Linear Systems via Output-Data-Driven Off-policy Q-learning Algorithm
    Zhang, Kun
    Liu, Xuantong
    Zhang, Lei
    Chen, Qian
    Peng, Yunjian
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2350 - 2356
  • [26] H∞ Filteing for Discrete-Time Markov Jump Linear Systems
    Che, Wei-Wei
    Guan, Wei
    2010 CHINESE CONTROL AND DECISION CONFERENCE, VOLS 1-5, 2010, : 350 - +
  • [27] Optimal tracking control for completely unknown nonlinear discrete-time Markov jump systems using data-based reinforcement learning method
    Jiang, He
    Zhang, Huaguang
    Luo, Yanhong
    Wang, Junyi
    NEUROCOMPUTING, 2016, 194 : 176 - 182
  • [28] Model-Free Q-Learning for the Tracking Problem of Linear Discrete-Time Systems
    Li, Chun
    Ding, Jinliang
    Lewis, Frank L.
    Chai, Tianyou
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3191 - 3201
  • [29] On Noncausal H∞ Tracking Control for Linear Discrete-Time Markovian Jump Systems
    Nakura, Gou
    2010 IEEE INTERNATIONAL CONFERENCE ON CONTROL APPLICATIONS, 2010, : 1981 - 1986
  • [30] An Optimal Tracking Control Method with Q-learning for Discrete-time Linear Switched System
    Zhao, Shangwei
    Wang, Jingcheng
    Wang, Hongyuan
    Xu, Haotian
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 1414 - 1419