Bridging Reinforcement Learning and Iterative Learning Control: Autonomous Motion Learning for Unknown, Nonlinear Dynamics

被引:12
作者
Meindl, Michael [1 ,2 ]
Lehmann, Dustin [3 ]
Seel, Thomas [2 ]
机构
[1] Hsch Karlsruhe, Embedded Mechatron Lab, Karlsruhe, Germany
[2] Friedrich Alexander Univ Erlangen Nurnberg, Dept Artificial Intelligence Biomed Engn, Erlangen, Germany
[3] Tech Univ Berlin, Control Syst Grp, Berlin, Germany
关键词
autonomous systems; Gaussian processes (GP); iterative learning control; nonlinear systems; reinforcement learning; robot learning; ADAPTIVE-CONTROL; DESIGN; FRAMEWORK; SYSTEMS; ROBOTS; TRADE;
D O I
10.3389/frobt.2022.793512
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
This work addresses the problem of reference tracking in autonomously learning robots with unknown, nonlinear dynamics. Existing solutions require model information or extensive parameter tuning, and have rarely been validated in real-world experiments. We propose a learning control scheme that learns to approximate the unknown dynamics by a Gaussian Process (GP), which is used to optimize and apply a feedforward control input on each trial. Unlike existing approaches, the proposed method neither requires knowledge of the system states and their dynamics nor knowledge of an effective feedback control structure. All algorithm parameters are chosen automatically, i.e. the learning method works plug and play. The proposed method is validated in extensive simulations and real-world experiments. In contrast to most existing work, we study learning dynamics for more than one motion task as well as the robustness of performance across a large range of learning parameters. The method's plug and play applicability is demonstrated by experiments with a balancing robot, in which the proposed method rapidly learns to track the desired output. Due to its model-agnostic and plug and play properties, the proposed method is expected to have high potential for application to a large class of reference tracking problems in systems with unknown, nonlinear dynamics.
引用
收藏
页数:14
相关论文
共 61 条
[1]   Iterative learning control: Brief survey and categorization [J].
Ahn, Hyo-Sung ;
Chen, YangQuan ;
Moore, Kevin L. .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2007, 37 (06) :1099-1121
[2]   High-Order Model-Free Adaptive Iterative Learning Control of Pneumatic Artificial Muscle With Enhanced Convergence [J].
Ai, Qingsong ;
Ke, Da ;
Zuo, Jie ;
Meng, Wei ;
Liu, Quan ;
Zhang, Zhiqiang ;
Xie, Sheng Q. .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2020, 67 (11) :9548-9559
[3]   Iterative learning control for discrete-time systems with exponential rate of convergence [J].
Amann, N ;
Owens, DH ;
Rogers, E .
IEE PROCEEDINGS-CONTROL THEORY AND APPLICATIONS, 1996, 143 (02) :217-224
[4]  
[Anonymous], 2010, EFFICIENT REINFORCEM
[5]  
Apgar T, 2018, ROBOTICS: SCIENCE AND SYSTEMS XIV
[6]   BETTERING OPERATION OF ROBOTS BY LEARNING [J].
ARIMOTO, S ;
KAWAMURA, S ;
MIYAZAKI, F .
JOURNAL OF ROBOTIC SYSTEMS, 1984, 1 (02) :123-140
[7]  
Atkeson CG, 1997, ARTIF INTELL REV, V11, P75, DOI 10.1023/A:1006511328852
[8]  
Berkenkamp F, 2015, 2015 EUROPEAN CONTROL CONFERENCE (ECC), P2496, DOI 10.1109/ECC.2015.7330913
[9]   A survey of iterative learning control [J].
Bristow, Douglas A. ;
Tharayil, Marina ;
Alleyne, Andrew G. .
IEEE CONTROL SYSTEMS MAGAZINE, 2006, 26 (03) :96-114
[10]  
Capone A, 2020, PR MACH LEARN RES, V120, P490