Learning from Physical Human Corrections, One Feature at a Time

被引:54
作者
Bajcsy, Andrea [1 ]
Losey, Dylan P. [2 ]
O'Malley, Marcia K. [2 ]
Dragan, Anca D. [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
[2] Rice Univ, Houston, TX 77251 USA
来源
HRI '18: PROCEEDINGS OF THE 2018 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION | 2018年
关键词
physical human-robot interaction; learning from demonstration; human teachers; MANIPULATION;
D O I
10.1145/3171221.3171267
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
We focus on learning robot objective functions from human guidance: specifically, from physical corrections provided by the person while the robot is acting. Objective functions are typically parametrized in terms of features, which capture aspects of the task that might be important. When the person intervenes to correct the robot's behavior, the robot should update its understanding of which features matter, how much, and in what way. Unfortunately, real users do not provide optimal corrections that isolate exactly what the robot was doing wrong. Thus, when receiving a correction, it is difficult for the robot to determine which features the person meant to correct, and which features were changed unintentionally. In this paper, we propose to improve the efficiency of robot learning during physical interactions by reducing unintended learning. Our approach allows the human-robot team to focus on learning one feature at a time, unlike state-of-the-art techniques that update all features at once. We derive an online method for identifying the single feature which the human is trying to change during physical interaction, and experimentally compare this one-at-a-time approach to the all-at-once baseline in a user study. Our results suggest that users teaching one-at-a-time perform better, especially in tasks that require changing multiple features.
引用
收藏
页码:141 / 149
页数:9
相关论文
共 24 条
  • [1] Keyframe-based Learning from Demonstration Method and Evaluation
    Akgun, Baris
    Cakmak, Maya
    Jiang, Karl
    Thomaz, Andrea L.
    [J]. INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2012, 4 (04) : 343 - 355
  • [2] [Anonymous], 2006, P 23 INT C MACH LEAR, DOI [10.1145/1143844.1143936, DOI 10.1145/1143844.1143936]
  • [3] [Anonymous], 2008, AAAI
  • [4] [Anonymous], 2007, URBANA
  • [5] [Anonymous], 2004, PROCEEDINGS OF THE T
  • [6] A survey of robot learning from demonstration
    Argall, Brenna D.
    Chernova, Sonia
    Veloso, Manuela
    Browning, Brett
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2009, 57 (05) : 469 - 483
  • [7] Bajcsy Andrea, 2017, C ROB LEARN CORL
  • [8] De Luca A, 2006, 2006 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-12, P1623
  • [9] An atlas of physical human-robot interaction
    De Santis, Agostino
    Siciliano, Bruno
    De Luca, Alessandro
    Bicchi, Antonio
    [J]. MECHANISM AND MACHINE THEORY, 2008, 43 (03) : 253 - 270
  • [10] Dragan AD, 2015, IEEE INT CONF ROBOT, P2339, DOI 10.1109/ICRA.2015.7139510