Parameter Optimal Iterative Learning Control Design: from Model-based, Data-driven to Reinforcement Learning *

被引:0
作者
Zhang, Yueqing [1 ]
Chu, Bing [1 ]
Shu, Zhan [2 ]
机构
[1] Univ Southampton, Southampton SO17 1BJ, Hants, England
[2] Univ Alberta, Edmonton, AB T6G 2H5, Canada
来源
IFAC PAPERSONLINE | 2022年 / 55卷 / 12期
关键词
Iterative learning control; reinforcement learning control; data-based control; TIME;
D O I
10.1016/j.ifacol.2022.07.360
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Iterative learning control (ILC) is a high-performance control design method for systems operating in a repetitive fashion by learning from past experience. Our recent work shows that reinforcement learning (RL) shares many features with ILC and thus opens the door to new ILC algorithm designs. This paper continues the research by considering a parameter optimal iterative learning control (POILC) algorithm. It has a very simple structure and appealing convergence properties, but requires a model of the system. We first develop a data-driven POILC algorithm without using model information by performing an extra experiment on the plant. We then use a policy gradient RL algorithm to design a new modelfree POILC algorithm. Both algorithms achieve the high-performance control target without using model information, but the convergence properties do differ. In particular, by increasing the number of function approximators in the latter, the RL-based model-free ILC can approach the performance of the model-based POILC. A numerical study is presented to compare the performance of different approaches and demonstrate the effectiveness of the proposed designs. Copyright 2022 The Authors. This is an open access article under the CC BY-NC-ND license (https://creativecommons.org/licenses/by-nc-nd/4.0/)
引用
收藏
页码:494 / 499
页数:6
相关论文
共 19 条
  • [1] Iterative learning control using optimal feedback and feedforward actions
    Amann, N
    Owens, DH
    Rogers, E
    [J]. INTERNATIONAL JOURNAL OF CONTROL, 1996, 65 (02) : 277 - 293
  • [2] Bertsekas D., 2019, COMPUTER SCI MATH
  • [3] Data-driven multivariable ILC: enhanced performance by eliminating L and Q filters
    Bolder, Joost
    Kleinendorst, Stephan
    Oomen, Tom
    [J]. INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2018, 28 (12) : 3728 - 3751
  • [4] A survey of iterative learning control
    Bristow, Douglas A.
    Tharayil, Marina
    Alleyne, Andrew G.
    [J]. IEEE CONTROL SYSTEMS MAGAZINE, 2006, 26 (03): : 96 - 114
  • [5] Reinforcement learning in continuous time and space
    Doya, K
    [J]. NEURAL COMPUTATION, 2000, 12 (01) : 219 - 245
  • [6] Discrete-time inverse model-based iterative learning control:: stability, monotonicity and robustness
    Harte, TJ
    Hätönen, J
    Owens, DH
    [J]. INTERNATIONAL JOURNAL OF CONTROL, 2005, 78 (08) : 577 - 586
  • [7] A Data-Driven Constrained Norm-Optimal Iterative Learning Control Framework for LTI Systems
    Janssens, Pieter
    Pipeleers, Goele
    Swevers, Jan
    [J]. IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2013, 21 (02) : 546 - 551
  • [8] An adaptive iterative learning control algorithm with experiments on an industrial robot
    Norrlöf, M
    [J]. IEEE TRANSACTIONS ON ROBOTICS AND AUTOMATION, 2002, 18 (02): : 245 - 251
  • [9] Robust monotone gradient-based discrete-time iterative learning control
    Owens, D. H.
    Hatonen, J. J.
    Daley, S.
    [J]. INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2009, 19 (06) : 634 - 661
  • [10] Owens DH, 2016, ADV IND CONTROL, P1, DOI 10.1007/978-1-4471-6772-3