Reinforcement Q-learning algorithm for H∞ tracking control of discrete-time Markov jump systems

被引:0
|
作者
Shi, Jiahui [1 ]
He, Dakuo [1 ,2 ]
Zhang, Qiang [1 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Shenyang, Peoples R China
[2] State Key Lab Synthet Automat Proc Ind, Shenyang, Peoples R China
基金
中国国家自然科学基金;
关键词
Markov jump systems; reinforcement learning; H-infinity tracking control; tracking game algebraicRiccati equation; NONLINEAR-SYSTEMS; DESIGN;
D O I
10.1080/00207721.2024.2395928
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, the H-infinity tracking control problem of linear discrete-time Markov jump systems is studied by using the data-based reinforcement learning method. Specifically, a new performance index function is established by using Markov chain and weighted sum technique, and thus the tracking game algebraic Riccati equation with weight vector and discount factor is obtained. A Q-learning algorithm is proposed to solve the tracking game algebra Riccati equation problem online without knowing the information of the system model. In addition, the convergence analysis of the algorithm is given, and it is proved that the added probing noise will not bias the algorithm. Finally, two simulation examples are given to verify the effectiveness of the proposed algorithm.
引用
收藏
页码:502 / 523
页数:22
相关论文
共 50 条
  • [31] Repetitive Control of Discrete-Time Markov Jump Linear Systems
    Ma, Guoqi
    Liu, Xinghua
    Pagilla, Prabhakar R.
    2018 ANNUAL AMERICAN CONTROL CONFERENCE (ACC), 2018, : 4546 - 4551
  • [32] Output-feedback Q-learning for discrete-time linear H∞ tracking control: A Stackelberg game approach
    Ren, Yunxiao
    Wang, Qishao
    Duan, Zhisheng
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2022, 32 (12) : 6805 - 6828
  • [33] Finite-Time H∞ Control for Discrete-Time Markov Jump Systems with Actuator Saturation
    Li, Bo
    Zhao, Junjie
    ABSTRACT AND APPLIED ANALYSIS, 2014,
  • [34] Reinforcement Q-Learning and Non-Zero-Sum Games Optimal Tracking Control for Discrete-Time Linear Multi-Input Systems
    Zhao, Jin-Gang
    2023 IEEE 12TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE, DDCLS, 2023, : 277 - 282
  • [35] Observer-Based H∞ Control on Nonhomogeneous Discrete-Time Markov Jump Systems
    Yin, Yanyan
    Shi, Peng
    Liu, Fei
    Teo, Kok Lay
    JOURNAL OF DYNAMIC SYSTEMS MEASUREMENT AND CONTROL-TRANSACTIONS OF THE ASME, 2013, 135 (04):
  • [36] Asynchronous iterative Q-learning based tracking control for nonlinear discrete-time multi-agent systems
    Shen, Ziwen
    Dong, Tao
    Huang, Tingwen
    NEURAL NETWORKS, 2024, 180
  • [37] Output Feedback Reinforcement Q-Learning Control for the Discrete-Time Linear Quadratic Regulator Problem
    Rizvi, Syed Ali Asad
    Lin, Zongli
    2017 IEEE 56TH ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2017,
  • [38] Improved Q-Learning Method for Linear Discrete-Time Systems
    Chen, Jian
    Wang, Jinhua
    Huang, Jie
    PROCESSES, 2020, 8 (03)
  • [39] Optimal Tracking Control for Linear Discrete-time Systems Using Reinforcement Learning
    Kiumarsi-Khomartash, Bahare
    Lewis, Frank L.
    Naghibi-Sistani, Mohammad-Bagher
    Karimpour, Ali
    2013 IEEE 52ND ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2013, : 3845 - 3850
  • [40] Zero-sum game-based optimal control for discrete-time Markov jump systems: A parallel off-policy Q-learning method
    Wang, Yun
    Fang, Tian
    Kong, Qingkai
    Li, Feng
    APPLIED MATHEMATICS AND COMPUTATION, 2024, 467