Reinforcement learning based compensation methods for robot manipulators

被引:72
作者
Pane, Yudha P. [1 ]
Nageshrao, Subramanya P. [2 ]
Kober, Jens [3 ]
Babuska, Robert [3 ]
机构
[1] Katholieke Univ Leuven, Dept Mech Engn, Div PMA, B-3001 Heverlee, Belgium
[2] Ford Motor Co, Green Field Lab, 3251 Hillview Ave, Palo Alto, CA 94304 USA
[3] Delft Univ Technol, Cognit Robot Dept, Mekelweg 2, NL-2628 CD Delft, Netherlands
关键词
Reinforcement learning; Tracking control; Robotics; Actor-critic scheme;
D O I
10.1016/j.engappai.2018.11.006
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Smart robotics will be a core feature while migrating from Industry 3.0 (i.e., mass manufacturing) to Industry 4.0 (i.e., customized or social manufacturing). A key characteristic of a smart system is its ability to learn. For smart manufacturing, this means incorporating learning capabilities into the current fixed, repetitive, task oriented industrial manipulators, thus rendering them 'smart'. In this paper we introduce two reinforcement learning (RL) based compensation methods. The learned correction signal, which compensates for unmodeled aberrations, is added to the existing nominal input with an objective to enhance the control performance. The proposed learning algorithms are evaluated on a 6-DoF industrial robotic manipulator arm to follow different kinds of reference paths, such as square or a circular path, or to track a trajectory on a three dimensional surface. In an extensive experimental study we compare the performance of our learning-based methods with well-known tracking controllers, namely, proportional-derivative (PD), model predictive control (MPC), and iterative learning control (ILC). The experimental results show a considerable performance improvement thanks to our RL-based methods when compared to PD, MPC, and ILC.
引用
收藏
页码:236 / 247
页数:12
相关论文
共 28 条
[1]  
An C. H., 1988, Model-Based Control of a Robot Manipulator
[2]  
[Anonymous], 2015, THESIS
[3]  
[Anonymous], 2003, P 3 IEEE RAS INT C H
[4]  
[Anonymous], 1998, Reinforcement learning: An introduction
[5]  
[Anonymous], 2015, IND 4 0 FUTURE PRODU
[6]  
Bayiz Y., 2014, 19 WORLD C INT FED A, P5393, DOI [DOI 10.3182/20140824-6-ZA-1003.02511, 10.3182/20140824-6-ZA-1003.02511]
[7]   A survey of iterative learning control [J].
Bristow, Douglas A. ;
Tharayil, Marina ;
Alleyne, Andrew G. .
IEEE CONTROL SYSTEMS MAGAZINE, 2006, 26 (03) :96-114
[8]   Reinforcement learning control of nonlinear multi-link system [J].
Bucak, IO ;
Zohdy, MA .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2001, 14 (05) :563-575
[9]  
Coates A, 2010, ENCY MACHINE LEARNIN, P53
[10]  
Conrad K.L., 2000, P 8 MED C CONTR AUT, V1719