H∞ Tracking Control for Linear Discrete-Time Systems: Model-Free Q-Learning Designs

被引:40
|
作者
Yang, Yunjie [1 ]
Wan, Yan [2 ]
Zhu, Jihong [1 ]
Lewis, Frank L. [3 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[2] Univ Texas Arlington, Dept Elect Engn, Arlington, TX 76019 USA
[3] Univ Texas Arlington, UTA Res Inst, Ft Worth, TX 75052 USA
来源
IEEE CONTROL SYSTEMS LETTERS | 2021年 / 5卷 / 01期
基金
中国国家自然科学基金;
关键词
Linear discrete-time systems; H-infinity tracking control; Q-learning; ZERO-SUM GAMES;
D O I
10.1109/LCSYS.2020.3001241
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this letter, a novel model-free Q-learning based approach is developed to solve the H-infinity tracking problem for linear discrete-time systems. A new exponential discounted value function is introduced that includes the cost of the whole control input and tracking error. The tracking Bellman equation and the game algebraic Riccati equation (GARE) are derived. The solution to the GARE leads to the feedback and feedforward parts of the control input. A Q-learning algorithm is then developed to learn the solution of the GARE online without requiring any knowledge of the system dynamics. Convergence of the algorithm is analyzed, and it is also proved that probing noises in maintaining the persistence of excitation (PE) condition do not result in any bias. An example of the F-16 aircraft short period dynamics is developed to validate the proposed algorithm.
引用
收藏
页码:175 / 180
页数:6
相关论文
共 50 条
  • [31] Output Feedback Q-Learning Control for the Discrete-Time Linear Quadratic Regulator Problem
    Rizvi, Syed Ali Asad
    Lin, Zongli
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (05) : 1523 - 1536
  • [32] Reinforcement Q-learning and Optimal Tracking Control of Unknown Discrete-time Multi-player Systems Based on Game Theory
    Zhao, Jin-Gang
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2024, 22 (05) : 1751 - 1759
  • [33] Output feedback Q-learning for discrete-time linear zero-sum games with application to the H-infinity control
    Rizvi, Syed Ali Asad
    Lin, Zongli
    AUTOMATICA, 2018, 95 : 213 - 221
  • [34] Data-Driven $H_{∞}$ Optimal Output Feedback Control for Linear Discrete-Time Systems Based on Off-Policy Q-Learning
    Zhang, Li
    Fan, Jialu
    Xue, Wenqian
    Lopez, Victor G.
    Li, Jinna
    Chai, Tianyou
    Lewis, Frank L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (07) : 3553 - 3567
  • [35] Model-Free Linear Discrete-Time System H∞ Control Using Input-Output Data
    Fan, Jialu
    Li, Zhen
    Jiang, Yi
    Chai, Tianyou
    Lewis, Frank L.
    2018 INTERNATIONAL CONFERENCE ON ADVANCED MECHATRONIC SYSTEMS (ICAMECHS), 2018, : 207 - 212
  • [36] Model-Free Global Stabilization of Discrete-time Linear Systems with Saturating Actuators Using Reinforcement Learning
    Rizvi, Syed Ali Asad
    Lin, Zongli
    2018 IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2018, : 5276 - 5281
  • [37] Reinforcement Q-Learning and Non-Zero-Sum Games Optimal Tracking Control for Discrete-Time Linear Multi-Input Systems
    Zhao, Jin-Gang
    2023 IEEE 12TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE, DDCLS, 2023, : 277 - 282
  • [38] Comparisons of Continuous-time and Discrete-time Q-learning Schemes for Adaptive Linear Quadratic Control
    Chun, Tae Yoon
    Lee, Jae Young
    Park, Jin Bae
    Choi, Yoon Ho
    2012 PROCEEDINGS OF SICE ANNUAL CONFERENCE (SICE), 2012, : 1228 - 1233
  • [39] H∞ Control for Discrete-Time Multi-Player Systems via Off-Policy Q-Learning
    Li, Jinna
    Xiao, Zhenfei
    IEEE ACCESS, 2020, 8 (08): : 28831 - 28846
  • [40] Output Feedback Reinforcement Q-Learning Control for the Discrete-Time Linear Quadratic Regulator Problem
    Rizvi, Syed Ali Asad
    Lin, Zongli
    2017 IEEE 56TH ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2017,