Learning Task Specifications from Demonstrations

被引:0
|
作者
Vazquez-Chanlatte, Marcell [1 ]
Jha, Susmit [2 ]
Tiwari, Ashish [2 ]
Ho, Mark K. [1 ]
Seshia, Sanjit A. [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
[2] SRI Int, 333 Ravenswood Ave, Menlo Pk, CA 94025 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018) | 2018年 / 31卷
基金
美国国家科学基金会;
关键词
MODEL;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Real-world applications often naturally decompose into several sub-tasks. In many settings (e.g., robotics) demonstrations provide a natural way to specify the sub-tasks. However, most methods for learning from demonstrations either do not provide guarantees that the artifacts learned for the sub-tasks can be safely recombined or limit the types of composition available. Motivated by this deficit, we consider the problem of inferring Boolean non-Markovian rewards (also known as logical trace properties or specifications) from demonstrations provided by an agent operating in an uncertain, stochastic environment. Crucially, specifications admit well-defined composition rules that are typically easy to interpret. In this paper, we formulate the specification inference task as a maximum a posteriori (MAP) probability inference problem, apply the principle of maximum entropy to derive an analytic demonstration likelihood model and give an efficient approach to search for the most likely specification in a large candidate pool of specifications. In our experiments, we demonstrate how learning specifications can help avoid common problems that often arise due to ad-hoc reward composition.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Learning from Corrective Demonstrations
    Gutierrez, Reymundo A.
    Short, Elaine Schaertl
    Niekum, Scott
    Thomaz, Andrea L.
    HRI '19: 2019 14TH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2019, : 712 - 714
  • [32] Robot learning of industrial assembly task via human demonstrations
    Maria Kyrarini
    Muhammad Abdul Haseeb
    Danijela Ristić-Durrant
    Axel Gräser
    Autonomous Robots, 2019, 43 : 239 - 257
  • [33] Shaping in reinforcement learning by knowledge transferred from human-demonstrations of a simple similar task
    Wang, Guo-Fang
    Fang, Zhou
    Li, Ping
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2018, 34 (01) : 711 - 720
  • [34] Learning From Visual Demonstrations via Replayed Task-Contrastive Model-Agnostic Meta-Learning
    Hu, Ziye
    Li, Wei
    Gan, Zhongxue
    Guo, Weikun
    Zhu, Jiwei
    Wen, James Zhiqing
    Zhou, Decheng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (12) : 8756 - 8767
  • [35] Elaborating on Learned Demonstrations with Temporal Logic Specifications
    Innes, Craig
    Ramamoorthy, Subramanian
    ROBOTICS: SCIENCE AND SYSTEMS XVI, 2020,
  • [36] Learning Options for an MDP from Demonstrations
    Tamassia, Marco
    Zambetta, Fabio
    Raffe, William
    Li, Xiaodong
    ARTIFICIAL LIFE AND COMPUTATIONAL INTELLIGENCE, 2015, 8955 : 226 - 242
  • [37] Robot Learning to Paint from Demonstrations
    Park, Younghyo
    Jeon, Seunghun
    Lee, Taeyoon
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 3053 - 3060
  • [38] Correspondence Learning Between Morphologically Different Robots via Task Demonstrations
    Aktas, Hakan
    Nagai, Yukie
    Asada, Minoru
    Oztop, Erhan
    Ugur, Emre
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (05) : 4463 - 4470
  • [39] Learning Task Specific Plans through Sound and Visually Interpretable Demonstrations
    Veeraraghavan, Harini
    Veloso, Manuela
    2008 IEEE/RSJ INTERNATIONAL CONFERENCE ON ROBOTS AND INTELLIGENT SYSTEMS, VOLS 1-3, CONFERENCE PROCEEDINGS, 2008, : 2599 - 2604
  • [40] Using physical demonstrations, background knowledge and vocal comments for task learning
    Pardowitz, M.
    Zoellner, R.
    Knoop, S.
    Dillmann, R.
    2006 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-12, 2006, : 322 - +