Model-Free control performance improvement using virtual reference feedback tuning and reinforcement Q-learning

被引:49
|
作者
Radac, Mircea-Bogdan [1 ]
Precup, Radu-Emil [1 ,2 ]
Roman, Raul-Cristian [1 ]
机构
[1] Politehn Univ Timisoara, Dept Automat & Appl Informat, Timisoara, Romania
[2] Edith Cowan Univ, Sch Engn, Joondalup, WA, Australia
关键词
Aerodynamic system; data-driven control; model-free control; position control; reinforcement Q-learning; virtual reference feedback tuning; CONTROL DESIGN; EXPERIMENTAL VALIDATION; TRAJECTORY TRACKING; SEARCH ALGORITHM; VRFT APPROACH; SYSTEMS; OPTIMIZATION; TORQUE;
D O I
10.1080/00207721.2016.1236423
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper proposes the combination of two model-free controller tuning techniques, namely linear virtual reference feedback tuning (VRFT) and nonlinear state-feedback Q-learning, referred to as a newmixed VRFT-Q learning approach. VRFT is first used to find stabilising feedback controller using input-output experimental data from the process in a model reference tracking setting. Reinforcement Q-learning is next applied in the same setting using input-state experimental data collected under perturbed VRFT to ensure good exploration. The Q-learning controller learned with a batch fitted Q iteration algorithm uses two neural networks, one for the Q-function estimator and one for the controller, respectively. The VRFT-Q learning approach is validated on position control of a two-degrees-of-motion open-loop stable multi input-multi output (MIMO) aerodynamic system (AS). Extensive simulations for the two independent control channels of theMIMO AS show that the Q-learning controllers clearly improve performance over the VRFT controllers.
引用
收藏
页码:1071 / 1083
页数:13
相关论文
共 50 条
  • [21] Model-free Control Design Using Policy Gradient Reinforcement Learning in LPV Framework
    Bao, Yajie
    Velni, Javad Mohammadpour
    2021 EUROPEAN CONTROL CONFERENCE (ECC), 2021, : 150 - 155
  • [22] Model-Free Control for Dynamic-Field Acoustic Manipulation Using Reinforcement Learning
    Latifi, Kourosh
    Kopitca, Artur
    Zhou, Quan
    IEEE ACCESS, 2020, 8 : 20597 - 20606
  • [23] Comparative study of model-based and model-free reinforcement learning control performance in HVAC systems
    Gao, Cheng
    Wang, Dan
    JOURNAL OF BUILDING ENGINEERING, 2023, 74
  • [24] Gain scheduled control of IPMC actuators with 'model-free' iterative feedback tuning
    McDaid, A. J.
    Aw, K. C.
    Xie, S. Q.
    Haemmerle, E.
    SENSORS AND ACTUATORS A-PHYSICAL, 2010, 164 (1-2) : 137 - 147
  • [25] Using Reinforcement Learning for Model-free Linear Quadratic Control with Process and Measurement Noises
    Yaghmaie, Farnaz Adib
    Gustafsson, Fredrik
    2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 6510 - 6517
  • [26] Heuristic action execution for energy efficient charge-sustaining control of connected hybrid vehicles with model-free double Q-learning
    Shuai, Bin
    Zhou, Quan
    Li, Ji
    He, Yinglong
    Li, Ziyang
    Williams, Huw
    Xu, Hongming
    Shuai, Shijin
    APPLIED ENERGY, 2020, 267
  • [27] Remarks on input to state stability of perturbed gradient flows, motivated by model-free feedback control learning
    Sontag, Eduardo D.
    SYSTEMS & CONTROL LETTERS, 2022, 161
  • [28] Fusion of Microgrid Control With Model-Free Reinforcement Learning: Review and Vision
    She, Buxin
    Li, Fangxing
    Cui, Hantao
    Zhang, Jingqiu
    Bo, Rui
    IEEE TRANSACTIONS ON SMART GRID, 2023, 14 (04) : 3232 - 3245
  • [29] Model-Free Reinforcement Learning of Minimal-Cost Variance Control
    Jing, Gangshan
    Bai, He
    George, Jemin
    Chakrabortty, Aranya
    IEEE CONTROL SYSTEMS LETTERS, 2020, 4 (04): : 916 - 921
  • [30] Model-Free H ∞ Output Feedback Control of Road Sensing in Vehicle Active Suspension Based on Reinforcement Learning
    Wang, Gang
    Li, Kunpeng
    Liu, Suqi
    Jing, Hui
    JOURNAL OF DYNAMIC SYSTEMS MEASUREMENT AND CONTROL-TRANSACTIONS OF THE ASME, 2023, 145 (06):