Influence Function Based Off-policy Q-learning Control for Markov Jump Systems

被引:0
|
作者
Yuling Zou [1 ]
Jiwei Wen [1 ]
Huiwen Xue [1 ]
Xiaoli Luan [1 ]
机构
[1] Jiangnan University,Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), the School of Internet of Things Engineering
关键词
control; influence function; Markov jump systems; off-policy Q-learning;
D O I
10.1007/s12555-024-0579-8
中图分类号
学科分类号
摘要
This paper presents an off-policy Q-learning approach based on influence function for addressing H∞ control of Markov jump systems. Unlike existing literatures, the mode classification and parallel update method is developed to directly decouple the relationship among matrices across different modes, tackling the most challenging aspect of this issue. Subsequently, we utilize the off-policy algorithm to derive the optimal policy, which allows for efficient learning without the need to follow the current policy being improved. This approach is particularly advantageous as it enables the algorithm to explore and evaluate different policies from historical data, thus circumventing the limitations associated with specific forms of disturbance updates. Moreover, the influence function is employed for data cleansing during the learning process, thereby enabling a more efficient learning period. A numerical example and a DC motor model are presented to illustrate the validity of the proposed method.
引用
收藏
页码:1411 / 1420
页数:9
相关论文
共 50 条
  • [21] Asynchronous Observer-Based Control for Exponential Stabilization of Markov Jump Systems
    Zhang, Meng
    Shen, Chao
    Wu, Zheng-Guang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2020, 67 (10) : 2039 - 2043
  • [22] Proportional integral observer based tracking control design for Markov jump systems
    Vijayakumar, M.
    Sakthivel, R.
    Mohammadzadeh, Ardashir
    Karthick, S. A.
    Anthoni, S. Marshal
    APPLIED MATHEMATICS AND COMPUTATION, 2021, 410 (410)
  • [23] Observer-based quantized sliding mode control of Markov jump systems
    Shen, Mouquan
    Zhang, Hainan
    Park, Ju H.
    NONLINEAR DYNAMICS, 2018, 92 (02) : 415 - 427
  • [24] Feedback Predictive Control Based on Periodic Invariant Set for Markov Jump Systems
    Cheng, Jingong
    Liu, Fei
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2015, 34 (08) : 2681 - 2693
  • [25] Observer-Based Control of 2-D Markov Jump Systems
    Le Van Hien
    Trinh, Hieu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2017, 64 (11) : 1322 - 1326
  • [26] Feedback Predictive Control Based on Periodic Invariant Set for Markov Jump Systems
    Jingong Cheng
    Fei Liu
    Circuits, Systems, and Signal Processing, 2015, 34 : 2681 - 2693
  • [27] Model-free adaptive optimal control policy for Markov jump systems: A value iterations algorithm
    Zhou, Peixin
    Wen, Jiwei
    Swain, Akshya Kumar
    Luan, Xiaoli
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART I-JOURNAL OF SYSTEMS AND CONTROL ENGINEERING, 2022, 236 (10) : 1819 - 1829
  • [28] Optimal H∞ control of Markov jump systems based on parallel Kleinman iteration algorithm
    Song J.
    He S.-P.
    Kongzhi yu Juece/Control and Decision, 2016, 31 (03): : 559 - 563
  • [29] Higher order moment stability region for Markov jump systems based on cumulant generating function
    Luan, Xiaoli
    Huang, Biao
    Liu, Fei
    AUTOMATICA, 2018, 93 : 389 - 396
  • [30] Fault tolerant control for networked Markov jump systems with data packet dropout based on observer
    Xu, Yating
    Sun, Xiaoyue
    Wang, Yanfeng
    Wang, Peiliang
    Li, Zuxin
    Xu, Hongyi
    PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 3727 - 3731