Robust H8 tracking of linear discrete-time systems using Q-learning

被引:2
作者
Valadbeigi, Amir Parviz [1 ,3 ]
Shu, Zhan [1 ]
Khaki Sedigh, Ali [2 ]
机构
[1] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB, Canada
[2] K N Toosi Univ Technol, Dept Elect Engn, Tehran, Iran
[3] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB T6G 1H9, Canada
关键词
auxiliary system; discounted factor; Q-learning; robust H-infinity tracking; H-INFINITY-CONTROL; ZERO-SUM GAMES; FEEDBACK-CONTROL; STABILIZATION; SYNCHRONIZATION;
D O I
10.1002/rnc.6662
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper deals with a robust H-infinity tracking problem with a discounted factor. A new auxiliary system is established in terms of norm-bounded time-varying uncertainties. It is shown that the robust discounted H-infinity tracking problem for the auxiliary system solves the original problem. Then, the new robust discounted H-infinity tracking problem is represented as a well-known zero-sum game problem. Moreover, the robust tracking Bellman equation and the robust tracking Algebraic Riccati equation (RTARE) are inferred. A lower bound of a discounted factor for stability is obtained to assure the stability of the closed-loop system. Based on the auxiliary system, the system is reshaped in a new structure that is applicable to Reinforcement Learning methods. Finally, an online Q-learning algorithm without the knowledge of system matrices is proposed to solve the algebraic Riccati equation associated with the robust discounted H-infinity tracking problem for the auxiliary system. Simulation results are given to verify the effectiveness and merits of the proposed method.
引用
收藏
页码:5604 / 5623
页数:20
相关论文
共 50 条
  • [31] Reinforcement Q-learning and Optimal Tracking Control of Unknown Discrete-time Multi-player Systems Based on Game Theory
    Zhao, Jin-Gang
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2024, 22 (05) : 1751 - 1759
  • [32] Continuous deep Q-learning with a simulator for stabilization of uncertain discrete-time systems
    Ikemoto, Junya
    Ushio, Toshimitsu
    IEICE NONLINEAR THEORY AND ITS APPLICATIONS, 2021, 12 (04): : 738 - 757
  • [33] Inverse Reinforcement Q-Learning Through Expert Imitation for Discrete-Time Systems
    Xue, Wenqian
    Lian, Bosen
    Fan, Jialu
    Kolaric, Patrik
    Chai, Tianyou
    Lewis, Frank L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (05) : 2386 - 2399
  • [34] Robust Inverse Q-Learning for Continuous-Time Linear Systems in Adversarial Environments
    Lian, Bosen
    Xue, Wenqian
    Lewis, Frank L.
    Chai, Tianyou
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (12) : 13083 - 13095
  • [35] Q-learning algorithm in solving consensusability problem of discrete-time multi-agent systems
    Feng, Tao
    Zhang, Jilie
    Tong, Yin
    Zhang, Huaguang
    AUTOMATICA, 2021, 128
  • [36] Finite-horizon Q-learning for discrete-time zero-sum games with application to H∞$$ {H}_{\infty } $$ control
    Liu, Mingxiang
    Cai, Qianqian
    Meng, Wei
    Li, Dandan
    Fu, Minyue
    ASIAN JOURNAL OF CONTROL, 2023, 25 (04) : 3160 - 3168
  • [37] Output Feedback Q-Learning Control for the Discrete-Time Linear Quadratic Regulator Problem
    Rizvi, Syed Ali Asad
    Lin, Zongli
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (05) : 1523 - 1536
  • [38] Reinforcement Q-Learning and Non-Zero-Sum Games Optimal Tracking Control for Discrete-Time Linear Multi-Input Systems
    Zhao, Jin-Gang
    2023 IEEE 12TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE, DDCLS, 2023, : 277 - 282
  • [39] Discrete-Time Multi-Player Games Based on Off-Policy Q-Learning
    Li, Jinna
    Xiao, Zhenfei
    Li, Ping
    IEEE ACCESS, 2019, 7 : 134647 - 134659
  • [40] Optimal tracking control for discrete-time systems by model-free off-policy Q-learning approach
    Li, Jinna
    Yuan, Decheng
    Ding, Zhengtao
    2017 11TH ASIAN CONTROL CONFERENCE (ASCC), 2017, : 7 - 12