H∞ Tracking learning control for discrete-time Markov jump systems: A parallel off-policy reinforcement learning

被引:1
|
作者
Zhang, Xuewen [1 ]
Xia, Jianwei [2 ]
Wang, Jing [1 ]
Chen, Xiangyong [3 ]
Shen, Hao [1 ]
机构
[1] Anhui Univ Technol, Sch Elect & Informat Engn, China Int Sci & Technol Cooperat Base Intelligent, Maanshan 243002, Peoples R China
[2] Liaocheng Univ, Sch Math Sci, Maanshan 252059, Peoples R China
[3] Linyi Univ, Sch Automat & Elect Engn, Linyi 276005, Peoples R China
来源
JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS | 2023年 / 360卷 / 18期
基金
中国国家自然科学基金;
关键词
FEEDBACK-CONTROL; LINEAR-SYSTEMS; DESIGN;
D O I
10.1016/j.jfranklin.2023.10.008
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper deals with the H infinity tracking control problem for a class of linear discrete-time Markov jump systems, in which the knowledge of system dynamics is not required. First, combined with reinforcement learning, a novel Bellman equation and the augmented coupled game algebraic Riccati equation are presented to derived the optimal control policy for the augmented discrete-time Markov jump systems. Moreover, based on the augmented system, a newly constructed system is given to collect the input and output data, which solves the problem that the coupling term in the discrete-time Markov jump systems is difficult to solve. Subsequently, a novel model-free algorithm is designed that does not need the dynamic information of the original system. Finally, a numerical example is given to verify the effectiveness of the proposed approach.
引用
收藏
页码:14878 / 14890
页数:13
相关论文
共 50 条
  • [31] Robust H∞ control of discrete-time Markov jump uncertain systems with bounded transition probabilities
    Haddad, Chafea
    Djennoune, Said
    Bettayeb, Maamar
    2013 10TH IEEE INTERNATIONAL CONFERENCE ON NETWORKING, SENSING AND CONTROL (ICNSC), 2013, : 473 - 478
  • [32] H∞ Static Output-Feedback Control Design for Discrete-Time Systems Using Reinforcement Learning
    Valadbeigi, Amir Parviz
    Sedigh, Ali Khaki
    Lewis, F. L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (02) : 396 - 406
  • [33] Off-Policy Reinforcement Learning: Optimal Operational Control for Two-Time-Scale Industrial Processes
    Li, Jinna
    Kiumarsi, Bahare
    Chai, Tianyou
    Lewis, Frank L.
    Fan, Jialu
    IEEE TRANSACTIONS ON CYBERNETICS, 2017, 47 (12) : 4547 - 4558
  • [34] Adaptive Fault-Tolerant Tracking Control for Discrete-Time Multiagent Systems via Reinforcement Learning Algorithm
    Li, Hongyi
    Wu, Ying
    Chen, Mou
    IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (03) : 1163 - 1174
  • [35] Robust Fault-tolerant Tracking Control for Linear Discrete-time Systems via Reinforcement Learning Method
    Nguyen, Ngoc Hoai An
    Kim, Sung Hyun
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2025, 23 (02) : 520 - 529
  • [36] H2-Filtering for discrete-time hidden Markov jump systems
    de Oliveira, A. M.
    Costa, O. L. V.
    INTERNATIONAL JOURNAL OF CONTROL, 2017, 90 (03) : 599 - 615
  • [37] H∞ filtering for discrete-time Markov jump systems with unknown transition probabilities
    Luan, Xiaoli
    Zhao, Shunyi
    Shi, Peng
    Liu, Fei
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2014, 28 (02) : 138 - 148
  • [38] H∞-Filtering Design for Discrete-Time Markov Jump Systems with Hidden Parameters
    de Oliveira, A. M.
    Costa, O. L. V.
    2016 IEEE CONFERENCE ON CONTROL APPLICATIONS (CCA), 2016,
  • [39] Event-based H∞ filtering for discrete-time Markov jump systems with network-induced delay
    Wang, Huijiao
    Xue, Anke
    Wang, Junhong
    Lu, Renquan
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2017, 354 (14): : 6170 - 6189
  • [40] Adaptive Fault-Tolerant Tracking Control for MIMO Discrete-Time Systems via Reinforcement Learning Algorithm With Less Learning Parameters
    Liu, Lei
    Wang, Zhanshan
    Zhang, Huaguang
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2017, 14 (01) : 299 - 313