Supervised autonomy for online learning in human-robot interaction

被引:30
作者
Senft, Emmanuel [1 ]
Baxter, Paul [2 ]
Kennedy, James [1 ]
Lemaignan, Severin [1 ]
Belpaeme, Tony [1 ,3 ]
机构
[1] Plymouth Univ, Plymouth PL4 8AA, Devon, England
[2] Univ Lincoln, Lincoln Ctr Autonomous Syst, Lincoln LN6 7TS, England
[3] Univ Ghent, Dept Elect & Informat Syst, Imec IDLab, Ghent, Belgium
关键词
Human-Robot interaction; Reinforcement learning; Interactive machine learning; Robotics; Progressive Autonomy; Supervised autonomy;
D O I
10.1016/j.patrec.2017.03.015
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When a robot is learning it needs to explore its environment and how its environment responds on its actions. When the environment is large and there are a large number of possible actions the robot can take, this exploration phase can take prohibitively long. However, exploration can often be optimised by letting a human expert guide the robot during its learning. Interactive machine learning, in which a human user interactively guides the robot as it learns, has been shown to be an effective way to teach a robot. It requires an intuitive control mechanism to allow the human expert to provide feedback on the robot's progress. This paper presents a novel method which combines Reinforcement Learning and Supervised Progressively Autonomous Robot Competencies (SPARC). By allowing the user to fully control the robot and by treating rewards as implicit, SPARC aims to learn an action policy while maintaining human supervisory oversight of the robot's behaviour. This method is evaluated and compared to Interactive Reinforcement Learning in a robot teaching task. Qualitative and quantitative results indicate that SPARC allows for safer and faster learning by the robot, whilst not placing a high workload on the human teacher. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:77 / 86
页数:10
相关论文
共 50 条
  • [11] Collaborative Autonomy: Human-Robot Interaction to the Test of Intelligent Help
    Cantucci, Filippo
    Falcone, Rino
    ELECTRONICS, 2022, 11 (19)
  • [12] Incremental learning of gestures for human-robot interaction
    Okada, Shogo
    Kobayashi, Yoichi
    Ishibashi, Satoshi
    Nishida, Toyoaki
    AI & SOCIETY, 2010, 25 (02) : 155 - 168
  • [13] Learning and Comfort in Human-Robot Interaction: A Review
    Wang, Weitian
    Chen, Yi
    Li, Rui
    Jia, Yunyi
    APPLIED SCIENCES-BASEL, 2019, 9 (23):
  • [14] Causal-HRI: Causal Learning for Human-Robot Interaction
    Cheong, Jiaee
    Churamani, Nikhil
    Guerdan, Luke
    Lee, Tabitha Edith
    Han, Zhao
    Gunes, Hatice
    COMPANION OF THE 2024 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2024 COMPANION, 2024, : 1311 - 1313
  • [15] Affordance-Based Human-Robot Interaction With Reinforcement Learning
    Munguia-Galeano, Francisco
    Veeramani, Satheeshkumar
    Hernandez, Juan David
    Wen, Qingmeng
    Ji, Ze
    IEEE ACCESS, 2023, 11 : 31282 - 31292
  • [16] A Hybrid PAC Reinforcement Learning Algorithm for Human-Robot Interaction
    Zehfroosh, Ashkan
    Tanner, Herbert G.
    FRONTIERS IN ROBOTICS AND AI, 2022, 9
  • [17] A survey of communicating robot learning during human-robot interaction
    Habibian, Soheil
    Valdivia, Antonio Alvarez
    Blumenschein, Laura H.
    Losey, Dylan P.
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2025, 44 (04) : 665 - 698
  • [18] Interaction Task Motion Learning for Human-Robot Interaction Control
    Lyu, Shangke
    Selvaraj, Nithish Muthuchamy
    Cheah, Chien Chern
    IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2022, 52 (05) : 894 - 906
  • [19] Human-robot interaction
    Murphy R.R.
    Nomura T.
    Billard A.
    Burke J.L.
    IEEE Robotics and Automation Magazine, 2010, 17 (02) : 85 - 89
  • [20] Online Assessment of Human-Robot Interaction for Hybrid Control of Walking
    del-Ama, Antonio J.
    Moreno, Juan C.
    Gil-Agudo, Angel
    de-los-Reyes, Ana
    Pons, Jose L.
    SENSORS, 2012, 12 (01) : 215 - 225