Influence Function Based Off-policy Q-learning Control for Markov Jump Systems

被引:0
|
作者
Yuling Zou [1 ]
Jiwei Wen [1 ]
Huiwen Xue [1 ]
Xiaoli Luan [1 ]
机构
[1] Jiangnan University,Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), the School of Internet of Things Engineering
关键词
control; influence function; Markov jump systems; off-policy Q-learning;
D O I
10.1007/s12555-024-0579-8
中图分类号
学科分类号
摘要
This paper presents an off-policy Q-learning approach based on influence function for addressing H∞ control of Markov jump systems. Unlike existing literatures, the mode classification and parallel update method is developed to directly decouple the relationship among matrices across different modes, tackling the most challenging aspect of this issue. Subsequently, we utilize the off-policy algorithm to derive the optimal policy, which allows for efficient learning without the need to follow the current policy being improved. This approach is particularly advantageous as it enables the algorithm to explore and evaluate different policies from historical data, thus circumventing the limitations associated with specific forms of disturbance updates. Moreover, the influence function is employed for data cleansing during the learning process, thereby enabling a more efficient learning period. A numerical example and a DC motor model are presented to illustrate the validity of the proposed method.
引用
收藏
页码:1411 / 1420
页数:9
相关论文
共 50 条
  • [31] Fuzzy-Model-Based Nonfragile Guaranteed Cost Control of Nonlinear Markov Jump Systems
    Wu, Zheng-Guang
    Dong, Shanling
    Shi, Peng
    Su, Hongye
    Huang, Tingwen
    Lu, Renquan
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2017, 47 (08): : 2388 - 2397
  • [32] Sliding mode control of Markov jump systems: Multi-processor based observer approach
    Zhang, Xiang
    He, Shuping
    Tan, Ying
    Zhang, Weidong
    AUTOMATICA, 2025, 174
  • [33] Disturbance-observer-based control for Markov jump systems with time-varying delay
    Gao, Qian
    Gao, Xianwen
    Qi, Wenhai
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2018, 39 (02) : 575 - 588
  • [34] Q-learning-based non-zero sum games for Markov jump multiplayer systems under actor-critic NNs structure
    Wang, Yun
    Xia, Jiawei
    Wang, Jing
    Shen, Hao
    INFORMATION SCIENCES, 2024, 681
  • [35] Reinforcement Learning Control of Double-Layer Markov Jump Systems With PDT-Switched Transition Probabilities
    Shen, Hao
    Wu, Jiacheng
    Wang, Yun
    Yan, Huaicheng
    Xu, Shengyuan
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024,
  • [36] Optimal control for continuous-time Markov jump singularly perturbed systems : A hybrid reinforcement learning scheme
    Huang, Yaling
    Li, Wenqian
    Wang, Yun
    Shen, Hao
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2024, 361 (07):
  • [37] Fuzzy-model-based robust control of Markov jump nonlinear systems with incomplete transition probabilities and uncertain packet dropouts
    Xu, Zeyuan
    Shi, Shaolong
    Er, Meng Joo
    Hu, Zhijian
    ASIAN JOURNAL OF CONTROL, 2023, 25 (06) : 4201 - 4214
  • [38] An intermediate-observer-based robust H 2 control of Markov jump systems with mismatched quantization
    Zhang, Hongjian
    Shen, Jie
    Shen, Mouquan
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2024, 361 (10):
  • [39] Network-based fuzzy control for nonlinear Markov jump systems subject to quantization and dropout compensation
    Zhang, Meng
    Shi, Peng
    Ma, Longhua
    Cai, Jianping
    Su, Hongye
    FUZZY SETS AND SYSTEMS, 2019, 371 : 96 - 109
  • [40] Dissipativity-based resilient asynchronous control for Markov jump systems with sector-bounded nonlinearities
    Zhang, Yujie
    Liu, Xiujuan
    Jiang, Junxia
    Xiao, Yongshan
    TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2018, 40 (09) : 2821 - 2830