Objective learning from human demonstrations

被引:6
|
作者
Lin, Jonathan Feng-Shun [1 ]
Carreno-Medrano, Pamela [2 ]
Parsapour, Mahsa [3 ]
Sakr, Maram [2 ,4 ]
Kulic, Dana [2 ]
机构
[1] Univ Waterloo, Syst Design Engn, Waterloo, ON, Canada
[2] Monash Univ, Fac Engn, Clayton, Vic, Australia
[3] Univ Waterloo, Elect & Comp Engn, Waterloo, ON, Canada
[4] Univ British Columbia, Mech Engn, Vancouver, BC, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Reward learning; Inverse optimal control; Inverse reinforcement learning; INVERSE OPTIMAL-CONTROL; COST-FUNCTIONS; GENERATION; ROBOT;
D O I
10.1016/j.arcontrol.2021.04.003
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Researchers in biomechanics, neuroscience, human-machine interaction and other fields are interested in inferring human intentions and objectives from observed actions. The problem of inferring objectives from observations has received extensive theoretical and methodological development from both the controls and machine learning communities. In this paper, we provide an integrating view of objective learning from human demonstration data. We differentiate algorithms based on the assumptions made about the objective function structure, how the similarity between the inferred objectives and the observed demonstrations is assessed, the assumptions made about the agent and environment model, and the properties of the observed human demonstrations. We review the application domains and validation approaches of existing works and identify the key open challenges and limitations. The paper concludes with an identification of promising directions for future work.
引用
收藏
页码:111 / 129
页数:19
相关论文
共 50 条
  • [1] Learning reward functions from diverse sources of human feedback: Optimally integrating demonstrations and preferences
    Biyik, Erdem
    Losey, Dylan P.
    Palan, Malayandi
    Landolfi, Nicholas C.
    Shevchuk, Gleb
    Sadigh, Dorsa
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2022, 41 (01): : 45 - 67
  • [2] Inverse KKT: Learning cost functions of manipulation tasks from demonstrations
    Englert, Peter
    Ngo Anh Vien
    Toussaint, Marc
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2017, 36 (13-14): : 1474 - 1488
  • [3] Learning Temporal Task Specifications From Demonstrations
    Baert, Mattijs
    Leroux, Sam
    Simoens, Pieter
    EXPLAINABLE AND TRANSPARENT AI AND MULTI-AGENT SYSTEMS, EXTRAAMAS 2024, 2024, 14847 : 81 - 98
  • [4] Learning Fairness from Demonstrations via Inverse Reinforcement Learning
    Blandin, Jack
    Kash, Ian
    PROCEEDINGS OF THE 2024 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, ACM FACCT 2024, 2024, : 51 - 61
  • [5] On stability for learning human control strategy by demonstrations using SVM
    Wang, Zhiyang
    Ou, Yongsheng
    ASSEMBLY AUTOMATION, 2020, 40 (01) : 118 - 131
  • [6] Unified Learning from Demonstrations, Corrections, and Preferences during Physical Human-Robot Interaction
    Mehta, Shaunak A.
    Losey, Dylan P.
    ACM TRANSACTIONS ON HUMAN-ROBOT INTERACTION, 2024, 13 (03)
  • [7] Learning Compliant Manipulation Tasks from Force Demonstrations
    Duan, Jianghua
    Ou, Yongsheng
    Xu, Sheng
    Wang, Zhiyang
    Peng, Ansi
    Wu, Xinyu
    Feng, Wei
    2018 IEEE INTERNATIONAL CONFERENCE ON CYBORG AND BIONIC SYSTEMS (CBS), 2018, : 449 - 454
  • [8] Learning to Serve: An Experimental Study for a New Learning From Demonstrations Framework
    Koc, Okan
    Peters, Jan
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (02) : 1784 - 1791
  • [9] Learning to Control Known Feedback Linearizable Systems From Demonstrations
    Sultangazin, Alimzhan
    Pannocchi, Luigi
    Fraile, Lucas
    Tabuada, Paulo
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (01) : 189 - 201
  • [10] Learning from Demonstrations for Autonomous Soft-tissue Retraction
    Pore, Ameya
    Tagliabue, Eleonora
    Piccinelli, Marco
    Dall'Alba, Diego
    Casals, Alicia
    Fiorini, Paolo
    2021 INTERNATIONAL SYMPOSIUM ON MEDICAL ROBOTICS (ISMR), 2021,