Data-Driven Model-Free Tracking Reinforcement Learning Control with VRFT-based Adaptive Actor-Critic

被引:35
作者
Radac, Mircea-Bogdan [1 ]
Precup, Radu-Emil [1 ]
机构
[1] Politehn Univ Timisoara, Dept Automat & Appl Informat, Timisoara 300006, Romania
来源
APPLIED SCIENCES-BASEL | 2019年 / 9卷 / 09期
关键词
adaptive actor-critic; model-free control; data-driven control; reinforcement learning; approximate dynamic programming; output reference model tracking; multi-input multi-output systems; vertical tank systems; Virtual Reference Feedback Tuning; CONJUGATE-GRADIENT ALGORITHM; TRAJECTORY TRACKING; NEURAL-NETWORKS; CONTROL DESIGN; SYSTEMS; FEEDBACK; ITERATION;
D O I
10.3390/app9091807
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
This paper proposes a neural network (NN)-based control scheme in an Adaptive Actor-Critic (AAC) learning framework designed for output reference model tracking, as a representative deep-learning application. The control learning scheme is model-free with respect to the process model. AAC designs usually require an initial controller to start the learning process; however, systematic guidelines for choosing the initial controller are not offered in the literature, especially in a model-free manner. Virtual Reference Feedback Tuning (VRFT) is proposed for obtaining an initially stabilizing NN nonlinear state-feedback controller, designed from input-state-output data collected from the process in open-loop setting. The solution offers systematic design guidelines for initial controller design. The resulting suboptimal state-feedback controller is next improved under the AAC learning framework by online adaptation of a critic NN and a controller NN. The mixed VRFT-AAC approach is validated on a multi-input multi-output nonlinear constrained coupled vertical two-tank system. Discussions on the control system behavior are offered together with comparisons with similar approaches.
引用
收藏
页数:24
相关论文
共 50 条
  • [41] An Actor-critic Reinforcement Learning Model for Optimal Bidding in Online Display Advertising
    Yuan, Congde
    Guo, Mengzhuo
    Xiang, Chaoneng
    Wang, Shuangyang
    Song, Guoqing
    Zhang, Qingpeng
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3604 - 3613
  • [42] An Actor-Critic Reinforcement Learning Control Approach for Discrete-Time Linear System with Uncertainty
    Chen, Hsin-Chang
    Lin, Yu-Chen
    Chang, Yu-Heng
    2018 INTERNATIONAL AUTOMATIC CONTROL CONFERENCE (CACS), 2018,
  • [43] Hierarchical Data-Driven Model-Free Iterative Learning Control Using Primitives
    Radac, Mircea-Bogdan
    Precup, Radu-Emil
    2016 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2016, : 2785 - 2790
  • [44] Improving Model-Free Control Algorithms Based on Data-Driven and Model-Driven Approaches: A Research Study
    Guo, Ziwei
    Yang, Huogen
    MATHEMATICS, 2024, 12 (01)
  • [45] Adaptive optimal coordination control of perturbed Bilateral Teleoperators with variable time delays using Actor-Critic Reinforcement Learning algorithm
    Dao, Phuong Nam
    Nguyen, Quang Phat
    Vu, Manh Hung
    MATHEMATICS AND COMPUTERS IN SIMULATION, 2025, 229 : 151 - 175
  • [46] Actor-Critic based Adaptive Control Strategy for Effective Energy Management
    Sankaranarayanan, Chandramouli
    Shaju, Sreenath
    Sukhwani, Mohak
    PROCEEDINGS OF THE 2022 5TH INTERNATIONAL CONFERENCE ON ADVANCED SYSTEMS AND EMERGENT TECHNOLOGIES IC_ASET'2022), 2022, : 23 - 28
  • [47] Model-free intelligent critic design with error analysis for neural tracking control
    Gao, Ning
    Wang, Ding
    Zhao, Mingming
    Hu, Lingzhi
    NEUROCOMPUTING, 2024, 572
  • [48] Model-Free Emergency Frequency Control Based on Reinforcement Learning
    Chen, Chunyu
    Cui, Mingjian
    Li, Fangxing
    Yin, Shengfei
    Wang, Xinan
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (04) : 2336 - 2346
  • [49] Model-Free control performance improvement using virtual reference feedback tuning and reinforcement Q-learning
    Radac, Mircea-Bogdan
    Precup, Radu-Emil
    Roman, Raul-Cristian
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2017, 48 (05) : 1071 - 1083
  • [50] Data-Driven Control of Hydraulic Manipulators by Reinforcement Learning
    Yao, Zhikai
    Xu, Fengyu
    Jiang, Guo-Ping
    Yao, Jianyong
    IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2024, 29 (04) : 2673 - 2684