Influence Function Based Off-policy Q-learning Control for Markov Jump Systems

被引:0
|
作者
Yuling Zou [1 ]
Jiwei Wen [1 ]
Huiwen Xue [1 ]
Xiaoli Luan [1 ]
机构
[1] Jiangnan University,Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), the School of Internet of Things Engineering
关键词
control; influence function; Markov jump systems; off-policy Q-learning;
D O I
10.1007/s12555-024-0579-8
中图分类号
学科分类号
摘要
This paper presents an off-policy Q-learning approach based on influence function for addressing H∞ control of Markov jump systems. Unlike existing literatures, the mode classification and parallel update method is developed to directly decouple the relationship among matrices across different modes, tackling the most challenging aspect of this issue. Subsequently, we utilize the off-policy algorithm to derive the optimal policy, which allows for efficient learning without the need to follow the current policy being improved. This approach is particularly advantageous as it enables the algorithm to explore and evaluate different policies from historical data, thus circumventing the limitations associated with specific forms of disturbance updates. Moreover, the influence function is employed for data cleansing during the learning process, thereby enabling a more efficient learning period. A numerical example and a DC motor model are presented to illustrate the validity of the proposed method.
引用
收藏
页码:1411 / 1420
页数:9
相关论文
共 50 条
  • [41] Compensation-Based Output Feedback Control for Fuzzy Markov Jump Systems With Random Packet Losses
    Xue, Min
    Yan, Huaicheng
    Zhang, Hao
    Zhan, Xisheng
    Shi, Kaibo
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (12) : 12759 - 12770
  • [42] Passivity-Based Control for Hidden Markov Jump Systems With Singular Perturbations and Partially Unknown Probabilities
    Li, Feng
    Xu, Shengyuan
    Shen, Hao
    Ma, Qian
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2020, 65 (08) : 3701 - 3706
  • [43] Reinforcement learning-based adaptive optimal tracking algorithm for Markov jump systems with partial unknown dynamics
    Tu, Yidong
    Fang, Haiyang
    Wang, Hai
    Shi, Kaibo
    He, Shuping
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2022, 43 (05) : 1435 - 1449
  • [44] Reinforcement Learning-Based Near Optimization for Continuous-Time Markov Jump Singularly Perturbed Systems
    Wang, Jing
    Peng, Chuanjun
    Park, Ju H.
    Shen, Hao
    Shi, Kaibo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (06) : 2026 - 2030
  • [45] Dynamic event-triggered control for discrete-time nonlinear Markov jump systems using policy iteration-based adaptive dynamic programming
    Tang, Fanghua
    Wang, Huanqing
    Chang, Xiao-Heng
    Zhang, Liang
    Alharbi, Khalid H.
    NONLINEAR ANALYSIS-HYBRID SYSTEMS, 2023, 49
  • [46] Optimal tracking control for completely unknown nonlinear discrete-time Markov jump systems using data-based reinforcement learning method
    Jiang, He
    Zhang, Huaguang
    Luo, Yanhong
    Wang, Junyi
    NEUROCOMPUTING, 2016, 194 : 176 - 182
  • [47] Dissipativity-based asynchronous control of discrete-time Markov jump systems with mixed time delays
    Zhang, Meng
    Shi, Peng
    Liu, Zhitao
    Cai, Jianping
    Su, Hongye
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2018, 28 (06) : 2161 - 2171
  • [48] Observer-based passive control of non-homogeneous Markov jump systems with random communication delays
    Chen, Yun
    Chen, Zhangping
    Chen, Zhenyu
    Xue, Anke
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2020, 51 (06) : 1133 - 1147
  • [49] Neural network-based adaptive reliable control for nonlinear Markov jump systems against actuator attacks
    Zhang, Junye
    Liu, Zhen
    Jiang, Baoping
    NONLINEAR DYNAMICS, 2023, 111 (15) : 13985 - 13999
  • [50] Neural network-based adaptive reliable control for nonlinear Markov jump systems against actuator attacks
    Junye Zhang
    Zhen Liu
    Baoping Jiang
    Nonlinear Dynamics, 2023, 111 : 13985 - 13999