Robust H8 tracking of linear discrete-time systems using Q-learning

被引:2
作者
Valadbeigi, Amir Parviz [1 ,3 ]
Shu, Zhan [1 ]
Khaki Sedigh, Ali [2 ]
机构
[1] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB, Canada
[2] K N Toosi Univ Technol, Dept Elect Engn, Tehran, Iran
[3] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB T6G 1H9, Canada
关键词
auxiliary system; discounted factor; Q-learning; robust H-infinity tracking; H-INFINITY-CONTROL; ZERO-SUM GAMES; FEEDBACK-CONTROL; STABILIZATION; SYNCHRONIZATION;
D O I
10.1002/rnc.6662
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper deals with a robust H-infinity tracking problem with a discounted factor. A new auxiliary system is established in terms of norm-bounded time-varying uncertainties. It is shown that the robust discounted H-infinity tracking problem for the auxiliary system solves the original problem. Then, the new robust discounted H-infinity tracking problem is represented as a well-known zero-sum game problem. Moreover, the robust tracking Bellman equation and the robust tracking Algebraic Riccati equation (RTARE) are inferred. A lower bound of a discounted factor for stability is obtained to assure the stability of the closed-loop system. Based on the auxiliary system, the system is reshaped in a new structure that is applicable to Reinforcement Learning methods. Finally, an online Q-learning algorithm without the knowledge of system matrices is proposed to solve the algebraic Riccati equation associated with the robust discounted H-infinity tracking problem for the auxiliary system. Simulation results are given to verify the effectiveness and merits of the proposed method.
引用
收藏
页码:5604 / 5623
页数:20
相关论文
共 50 条
  • [21] A non-fragile robust H8 regulator in discrete-time control systems with nonzero reference
    Ghaffari, Valiollah
    COMMUNICATIONS IN NONLINEAR SCIENCE AND NUMERICAL SIMULATION, 2023, 125
  • [22] Game Theoretical Reinforcement Learning for Robust H∞ Tracking Control of Discrete-Time Linear Systems with Unknown Dynamics
    Wu, Hao
    Li, Shaobao
    Durdevic, Petar
    Yang, Zhenyu
    2021 4TH INTERNATIONAL CONFERENCE ON INTELLIGENT AUTONOMOUS SYSTEMS (ICOIAS 2021), 2021, : 290 - 295
  • [23] FINITE-HORIZON OPTIMAL CONTROL OF DISCRETE-TIME LINEAR SYSTEMS WITH COMPLETELY UNKNOWN DYNAMICS USING Q-LEARNING
    Zhao, Jingang
    Zhang, Chi
    JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2021, 17 (03) : 1471 - 1483
  • [24] On bounded realness for H8 control design in discrete-time descriptor systems
    Park, Chan-eun
    Park, In Seok
    Kwon, Nam Kyu
    Park, PooGyeon
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (12) : 7246 - 7259
  • [25] Explainable data-driven Q-learning control for a class of discrete-time linear autonomous systems
    Perrusquia, Adolfo
    Zou, Mengbang
    Guo, Weisi
    INFORMATION SCIENCES, 2024, 682
  • [26] An iterative Q-learning scheme for the global stabilization of discrete-time linear systems subject to actuator saturation
    Rizvi, Syed Ali Asad
    Lin, Zongli
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2019, 29 (09) : 2660 - 2672
  • [27] Experience replay-based output feedback Q-learning scheme for optimal output tracking control of discrete-time linear systems
    Rizvi, Syed Ali Asad
    Lin, Zongli
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2019, 33 (12) : 1825 - 1842
  • [28] Q-learning solution for optimal consensus control of discrete-time multiagent systems using reinforcement learning
    Mu, Chaoxu
    Zhao, Qian
    Gao, Zhongke
    Sun, Changyin
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2019, 356 (13): : 6946 - 6967
  • [29] A DISCRETE-TIME SWITCHING SYSTEM ANALYSIS OF Q-LEARNING
    Lee, Donghwan
    Hu, Jianghai
    He, Niao
    SIAM JOURNAL ON CONTROL AND OPTIMIZATION, 2023, 61 (03) : 1861 - 1880
  • [30] Output Feedback H∞ Control for Linear Discrete-Time Multi-Player Systems With Multi-Source Disturbances Using Off-Policy Q-Learning
    Xiao, Zhenfei
    Li, Jinna
    Li, Ping
    IEEE ACCESS, 2020, 8 : 208938 - 208951