The BesMan Learning Platform for Automated Robot Skill Learning

被引:10
作者
Gutzeit, Lisa [1 ]
Fabisch, Alexander [2 ]
Otto, Marc [1 ]
Metzen, Jan Hendrik [3 ]
Hansen, Jonas [1 ]
Kirchner, Frank [1 ,2 ]
Kirchner, Elsa Andrea [1 ,2 ]
机构
[1] Univ Bremen, Robot Res Grp, Bremen, Germany
[2] DFKI GmbH, Robot Innovat Ctr, Bremen, Germany
[3] Robert Bosch GmbH, Bosch Ctr Artificial Intelligence, Renningen, Germany
关键词
robotics; imitiation learning; reinforcement learning; manipulation; behavior segmentation;
D O I
10.3389/frobt.2018.00043
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
We describe the BesMan learning platform which allows learning robotic manipulation behavior. It is a stand-alone solution which can be combined with different robotic systems and applications. Behavior that is adaptive to task changes and different target platforms can be learned to solve unforeseen challenges and tasks, which can occur during deployment of a robot. The learning platform is composed of components that deal with preprocessing of human demonstrations, segmenting the demonstrated behavior into basic building blocks, imitation, refinement by means of reinforcement learning, and generalization to related tasks. The core components are evaluated in an empirical study with 10 participants with respect to automation level and time requirements. We show that most of the required steps for transferring skills from humans to robots can be automated and all steps can be performed in reasonable time allowing to apply the learning platform on demand.
引用
收藏
页数:14
相关论文
共 57 条
[1]   A Shift in Task Routines During the Learning of a Motor Skill: Group-Averaged Data May Mask Critical Phases in the Individuals' Acquisition of Skilled Performance [J].
Adi-Japha, Esther ;
Karni, Avi ;
Parnes, Ariel ;
Loewenschuss, Iris ;
Vakil, Eli .
JOURNAL OF EXPERIMENTAL PSYCHOLOGY-LEARNING MEMORY AND COGNITION, 2008, 34 (06) :1544-1551
[2]   A survey of robot learning from demonstration [J].
Argall, Brenna D. ;
Chernova, Sonia ;
Veloso, Manuela ;
Browning, Brett .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2009, 57 (05) :469-483
[3]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[4]  
Atkeson S. S. C. G., 1997, ADV NEURAL INFORM PR, P12
[5]  
Bargsten V., 2015, 8 INT WORKSH HUM FRI
[6]  
Bartsch S, 2016, P ISR 2016 47 INT S, P1
[7]   Nonlinear system identification using coevolution of models and tests [J].
Bongard, JC ;
Lipson, H .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2005, 9 (04) :361-384
[8]  
Bongardt B., 2015, THESIS
[9]  
Brochu E., 2010, ARXIV10122599 CORR
[10]   Bayesian optimization for learning gaits under uncertainty [J].
Calandra, Roberto ;
Seyfarth, Andre ;
Peters, Jan ;
Deisenroth, Marc Peter .
ANNALS OF MATHEMATICS AND ARTIFICIAL INTELLIGENCE, 2016, 76 (1-2) :5-23