H∞ Tracking learning control for discrete-time Markov jump systems: A parallel off-policy reinforcement learning

被引:1
|
作者
Zhang, Xuewen [1 ]
Xia, Jianwei [2 ]
Wang, Jing [1 ]
Chen, Xiangyong [3 ]
Shen, Hao [1 ]
机构
[1] Anhui Univ Technol, Sch Elect & Informat Engn, China Int Sci & Technol Cooperat Base Intelligent, Maanshan 243002, Peoples R China
[2] Liaocheng Univ, Sch Math Sci, Maanshan 252059, Peoples R China
[3] Linyi Univ, Sch Automat & Elect Engn, Linyi 276005, Peoples R China
来源
JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS | 2023年 / 360卷 / 18期
基金
中国国家自然科学基金;
关键词
FEEDBACK-CONTROL; LINEAR-SYSTEMS; DESIGN;
D O I
10.1016/j.jfranklin.2023.10.008
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper deals with the H infinity tracking control problem for a class of linear discrete-time Markov jump systems, in which the knowledge of system dynamics is not required. First, combined with reinforcement learning, a novel Bellman equation and the augmented coupled game algebraic Riccati equation are presented to derived the optimal control policy for the augmented discrete-time Markov jump systems. Moreover, based on the augmented system, a newly constructed system is given to collect the input and output data, which solves the problem that the coupling term in the discrete-time Markov jump systems is difficult to solve. Subsequently, a novel model-free algorithm is designed that does not need the dynamic information of the original system. Finally, a numerical example is given to verify the effectiveness of the proposed approach.
引用
收藏
页码:14878 / 14890
页数:13
相关论文
共 50 条
  • [21] H∞ for discrete-time singular Markov jump systems subject to actuator saturation
    Ma, Shuping
    Zhang, Chenghui
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2012, 349 (03): : 1011 - 1029
  • [22] Observer-Based H∞ Control on Nonhomogeneous Discrete-Time Markov Jump Systems
    Yin, Yanyan
    Shi, Peng
    Liu, Fei
    Teo, Kok Lay
    JOURNAL OF DYNAMIC SYSTEMS MEASUREMENT AND CONTROL-TRANSACTIONS OF THE ASME, 2013, 135 (04):
  • [23] Output-feedback Quadratic Tracking Control of Continuous-time Systems by Using Off-policy Reinforcement Learning with Neural Networks Observer
    Meng, Qingqing
    Peng, Yunjian
    PROCEEDINGS OF THE 32ND 2020 CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2020), 2020, : 1504 - 1509
  • [24] Output Feedback H∞ Control for Linear Discrete-Time Multi-Player Systems With Multi-Source Disturbances Using Off-Policy Q-Learning
    Xiao, Zhenfei
    Li, Jinna
    Li, Ping
    IEEE ACCESS, 2020, 8 : 208938 - 208951
  • [25] Efficient Off-Policy Q-Learning for Data-Based Discrete-Time LQR Problems
    Lopez, Victor G.
    Alsalti, Mohammad
    Mueller, Matthias A.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (05) : 2922 - 2933
  • [26] Online optimal and adaptive integral tracking control for varying discrete-time systems using reinforcement learning
    Sanusi, Ibrahim
    Mills, Andrew
    Dodd, Tony
    Konstantopoulos, George
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2020, 34 (08) : 971 - 991
  • [27] Optimal Synchronization Control of Multiagent Systems With Input Saturation via Off-Policy Reinforcement Learning
    Qin, Jiahu
    Li, Man
    Shi, Yang
    Ma, Qichao
    Zheng, Wei Xing
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (01) : 85 - 96
  • [28] Finite-time asynchronous H filtering for discrete-time Markov jump systems over a lossy network
    Shen, Hao
    Li, Feng
    Wu, Zheng-Guang
    Park, Ju H.
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2016, 26 (17) : 3831 - 3848
  • [29] Stochastic optimal control problems of discrete-time Markov jump systems
    Song, Teng
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2023, 44 (05) : 2551 - 2570
  • [30] Robust H∞ Control for Discrete-Time Uncertain Descriptor Markov Jump Systems with Actuator Saturation
    Chen, Naixun
    Ma, Shuping
    PROCEEDINGS OF THE 31ST CHINESE CONTROL CONFERENCE, 2012, : 2662 - 2667