Robust H8 tracking of linear discrete-time systems using Q-learning

被引:2
作者
Valadbeigi, Amir Parviz [1 ,3 ]
Shu, Zhan [1 ]
Khaki Sedigh, Ali [2 ]
机构
[1] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB, Canada
[2] K N Toosi Univ Technol, Dept Elect Engn, Tehran, Iran
[3] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB T6G 1H9, Canada
关键词
auxiliary system; discounted factor; Q-learning; robust H-infinity tracking; H-INFINITY-CONTROL; ZERO-SUM GAMES; FEEDBACK-CONTROL; STABILIZATION; SYNCHRONIZATION;
D O I
10.1002/rnc.6662
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper deals with a robust H-infinity tracking problem with a discounted factor. A new auxiliary system is established in terms of norm-bounded time-varying uncertainties. It is shown that the robust discounted H-infinity tracking problem for the auxiliary system solves the original problem. Then, the new robust discounted H-infinity tracking problem is represented as a well-known zero-sum game problem. Moreover, the robust tracking Bellman equation and the robust tracking Algebraic Riccati equation (RTARE) are inferred. A lower bound of a discounted factor for stability is obtained to assure the stability of the closed-loop system. Based on the auxiliary system, the system is reshaped in a new structure that is applicable to Reinforcement Learning methods. Finally, an online Q-learning algorithm without the knowledge of system matrices is proposed to solve the algebraic Riccati equation associated with the robust discounted H-infinity tracking problem for the auxiliary system. Simulation results are given to verify the effectiveness and merits of the proposed method.
引用
收藏
页码:5604 / 5623
页数:20
相关论文
共 50 条
  • [41] Off-Policy Interleaved Q-Learning: Optimal Control for Affine Nonlinear Discrete-Time Systems
    Li, Jinna
    Chai, Tianyou
    Lewis, Frank L.
    Ding, Zhengtao
    Jiang, Yi
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (05) : 1308 - 1320
  • [42] A novel policy iteration based deterministic Q-learning for discrete-time nonlinear systems
    Wei QingLai
    Liu DeRong
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2015, 58 (12) : 1 - 15
  • [43] Comparisons of Continuous-time and Discrete-time Q-learning Schemes for Adaptive Linear Quadratic Control
    Chun, Tae Yoon
    Lee, Jae Young
    Park, Jin Bae
    Choi, Yoon Ho
    [J]. 2012 PROCEEDINGS OF SICE ANNUAL CONFERENCE (SICE), 2012, : 1228 - 1233
  • [44] A novel policy iteration based deterministic Q-learning for discrete-time nonlinear systems
    WEI QingLai
    LIU DeRong
    [J]. ScienceChina(InformationSciences), 2015, 58 (12) : 147 - 161
  • [45] Output Feedback Reinforcement Q-Learning Control for the Discrete-Time Linear Quadratic Regulator Problem
    Rizvi, Syed Ali Asad
    Lin, Zongli
    [J]. 2017 IEEE 56TH ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2017,
  • [46] Networked H∞ filtering for linear discrete-time systems
    Song, Hongbo
    Yu, Li
    Zhang, Wen-An
    [J]. INFORMATION SCIENCES, 2011, 181 (03) : 686 - 696
  • [47] Stochastic linear quadratic optimal control for model-free discrete-time systems based on Q-learning algorithm
    Wang, Tao
    Zhang, Huaguang
    Luo, Yanhong
    [J]. NEUROCOMPUTING, 2018, 312 : 1 - 8
  • [48] Model-free Q-learning designs for linear discrete-time zero-sum games with application to H-infinity control
    Al-Tamimi, Asma
    Lewis, Frank L.
    Abu-Khalaf, Murad
    [J]. AUTOMATICA, 2007, 43 (03) : 473 - 481
  • [49] Off-policy Q-learning-based Tracking Control for Stochastic Linear Discrete-Time Systems
    Liu, Xuantong
    Zhang, Lei
    Peng, Yunjian
    [J]. 2022 4TH INTERNATIONAL CONFERENCE ON CONTROL AND ROBOTICS, ICCR, 2022, : 252 - 256
  • [50] Neural Q-learning for discrete-time nonlinear zero-sum games with adjustable convergence rate
    Wang, Yuan
    Wang, Ding
    Zhao, Mingming
    Liu, Nan
    Qiao, Junfei
    [J]. NEURAL NETWORKS, 2024, 175