Objective learning from human demonstrations

被引:6
作者
Lin, Jonathan Feng-Shun [1 ]
Carreno-Medrano, Pamela [2 ]
Parsapour, Mahsa [3 ]
Sakr, Maram [2 ,4 ]
Kulic, Dana [2 ]
机构
[1] Univ Waterloo, Syst Design Engn, Waterloo, ON, Canada
[2] Monash Univ, Fac Engn, Clayton, Vic, Australia
[3] Univ Waterloo, Elect & Comp Engn, Waterloo, ON, Canada
[4] Univ British Columbia, Mech Engn, Vancouver, BC, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Reward learning; Inverse optimal control; Inverse reinforcement learning; INVERSE OPTIMAL-CONTROL; COST-FUNCTIONS; GENERATION; ROBOT;
D O I
10.1016/j.arcontrol.2021.04.003
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Researchers in biomechanics, neuroscience, human-machine interaction and other fields are interested in inferring human intentions and objectives from observed actions. The problem of inferring objectives from observations has received extensive theoretical and methodological development from both the controls and machine learning communities. In this paper, we provide an integrating view of objective learning from human demonstration data. We differentiate algorithms based on the assumptions made about the objective function structure, how the similarity between the inferred objectives and the observed demonstrations is assessed, the assumptions made about the agent and environment model, and the properties of the observed human demonstrations. We review the application domains and validation approaches of existing works and identify the key open challenges and limitations. The paper concludes with an identification of promising directions for future work.
引用
收藏
页码:111 / 129
页数:19
相关论文
共 50 条
  • [41] Social interaction for efficient agent learning from human reward
    Li, Guangliang
    Whiteson, Shimon
    Knox, W. Bradley
    Hung, Hayley
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2018, 32 (01) : 1 - 25
  • [42] Objective assessment of the human visual attentional state
    Willeford, Kevin T.
    Ciuffreda, Kenneth J.
    Yadav, Naveen K.
    Ludlam, Diana P.
    DOCUMENTA OPHTHALMOLOGICA, 2013, 126 (01) : 29 - 44
  • [43] Learning How to Drive in Blind Intersections from Human Data
    Sama, Kyle
    Morales, Yoichi
    Akai, Naoki
    Takeuchi, Eijiro
    Takeda, Kazuya
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 317 - 324
  • [44] Interactive Reinforcement Learning from Demonstration and Human Evaluative Feedback
    Li, Guangliang
    Gomez, Randy
    Nakamura, Keisuke
    He, Bo
    2018 27TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (IEEE RO-MAN 2018), 2018, : 1156 - 1162
  • [45] Learning from Richer Human Guidance: Augmenting Comparison-Based Learning with Feature Queries
    Basu, Chandrayee
    Singhal, Mukesh
    Dragan, Anca D.
    HRI '18: PROCEEDINGS OF THE 2018 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2018, : 132 - 140
  • [46] Neural Task Success Classifiers for Robotic Manipulation from Few Real Demonstrations
    Mohtasib, Abdalkarim
    Ghalamzan, Amir E.
    Bellotto, Nicola
    Cuayahuitl, Heriberto
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [47] Condition-based Design of Variable Impedance Controllers from User Demonstrations
    San-Miguel, Alberto
    Puig, Vicenc
    Alenya, Guillem
    CONTROL ENGINEERING PRACTICE, 2023, 140
  • [48] Human-Automation Interaction for Assisting Novices to Emulate Experts by Inferring Task Objective Functions
    Byeon, Sooyung
    Jin, Wanxin
    Sun, Dawei
    Hwang, Inseok
    2021 IEEE/AIAA 40TH DIGITAL AVIONICS SYSTEMS CONFERENCE (DASC), 2021,
  • [49] Synthesizing human brain CT images from MRI with deep learning
    Zhou, Longfei
    Pu, Chenghan
    Lu, Yuyang
    Zhang, Ying
    Han, Ke
    Jin, YiRui
    Zhang, Shengquan
    Liu, Kaihua
    INTERNATIONAL JOURNAL OF MODELING SIMULATION AND SCIENTIFIC COMPUTING, 2024, 15 (01)
  • [50] Learning Task-Oriented Dexterous Grasping from Human Knowledge
    Li, Hui
    Zhang, Yinlong
    Li, Yanan
    He, Hongsheng
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 6192 - 6198