Q-learning based tracking control with novel finite-horizon performance index ☆

被引:0
|
作者
Wang, Wei [1 ,2 ,3 ]
Wang, Ke [1 ]
Huang, Zixin [4 ]
Mu, Chaoxu [1 ]
Shi, Haoxian [5 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
[2] Zhongnan Univ Econ & Law, Sch Informat Engn, Wuhan 430073, Peoples R China
[3] Zhongnan Univ Econ & Law, Emergency Management Res Ctr, Wuhan 430073, Peoples R China
[4] Wuhan Inst Technol, Sch Elect & Informat Engn, Wuhan 430205, Peoples R China
[5] China Geol Survey, Guangzhou Marine Geol Survey, Guangzhou 510075, Peoples R China
关键词
Optimal tracking control; Model-free control; Q-function; Finite-horizon; NONLINEAR-SYSTEMS; TIME-SYSTEMS;
D O I
10.1016/j.ins.2024.121212
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A data-driven method is designed to realize the model-free finite-horizon optimal tracking control (FHOTC) of unknown linear discrete-time systems based on Q-learning in this paper. First, a novel finite-horizon performance index (FHPI) that only depends on the next-step tracking error is introduced. Then, an augmented system is formulated, which incorporates with the system model and the trajectory model. Based on the novel FHPI, a derivation of the augmented time-varying Riccati equation (ATVRE) is provided. We present a data-driven FHOTC method that uses Qlearning to optimize the defined time-varying Q-function. This allows us to estimate the solutions of the ATVRE without the system dynamics. Finally, the validity and features of the proposed Qlearning-based FHOTC method are demonstrated by means of conducting comparative simulation studies.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] FINITE-HORIZON OPTIMAL CONTROL OF DISCRETE-TIME LINEAR SYSTEMS WITH COMPLETELY UNKNOWN DYNAMICS USING Q-LEARNING
    Zhao, Jingang
    Zhang, Chi
    JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2021, 17 (03) : 1471 - 1483
  • [2] Finite-horizon H∞ tracking control for discrete-time linear systems
    Wang, Jian
    Wang, Wei
    Liang, Xiaofeng
    Zuo, Chao
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2024, 34 (01) : 54 - 70
  • [3] Discounted linear Q-learning control with novel tracking cost and its stability
    Wang, Ding
    Ren, Jin
    Ha, Mingming
    INFORMATION SCIENCES, 2023, 626 : 339 - 353
  • [4] Finite-horizon Q-learning for discrete-time zero-sum games with application to H∞$$ {H}_{\infty } $$ control
    Liu, Mingxiang
    Cai, Qianqian
    Meng, Wei
    Li, Dandan
    Fu, Minyue
    ASIAN JOURNAL OF CONTROL, 2023, 25 (04) : 3160 - 3168
  • [5] Output feedback Q-learning for discrete-time finite-horizon zero-sum games with application to the H? control
    Liu, Mingxiang
    Cai, Qianqian
    Li, Dandan
    Meng, Wei
    Fu, Minyue
    NEUROCOMPUTING, 2023, 529 : 48 - 55
  • [6] Analyses for Optimal Control of Discrete Time-Delay Systems Based on ADP Algorithm with Finite-Horizon Performance Index
    Song, Ruizhuo
    Xing, Shi
    PROCEEDINGS OF THE 28TH CHINESE CONTROL AND DECISION CONFERENCE (2016 CCDC), 2016, : 7076 - 7080
  • [7] Finite-Horizon Discounted Optimal Control: Stability and Performance
    Granzotto, Mathieu
    Postoyan, Romain
    Busoniu, Lucian
    Nesic, Dragan
    Daafouz, Jamal
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2021, 66 (02) : 550 - 565
  • [8] Finite-horizon optimal secure tracking control under denial-of-service attacks
    Wang, Jian
    Wang, Wei
    Liang, Xiaofeng
    ISA TRANSACTIONS, 2024, 149 : 44 - 53
  • [9] Computing a Classic Index for Finite-Horizon Bandits
    Nino-Mora, Jose
    INFORMS JOURNAL ON COMPUTING, 2011, 23 (02) : 254 - 267
  • [10] Approximate Finite-horizon Optimal Control with Policy Iteration
    Zhao Zhengen
    Yang Ying
    Li Hao
    Liu Dan
    2014 33RD CHINESE CONTROL CONFERENCE (CCC), 2014, : 8889 - 8894