Transfer in Inverse Reinforcement Learning for Multiple Strategies

被引:0
|
作者
Tanwani, Ajay Kumar [1 ]
Billard, Aude [1 ]
机构
[1] Ecole Polytech Fed Lausanne, LASA, CH-1015 Lausanne, Switzerland
来源
2013 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS) | 2013年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the problem of incrementally learning different strategies of performing a complex sequential task from multiple demonstrations of an expert or a set of experts. While the task is the same, each expert differs in his/her way of performing it. We assume that this variety across experts' demonstration is due to the fact that each expert/strategy is driven by a different reward function, where reward function is expressed as a linear combination of a set of known features. Consequently, we can learn all the expert strategies by forming a convex set of optimal deterministic policies, from which one can match any unseen expert strategy drawn from this set. Instead of learning from scratch every optimal policy in this set, the learner transfers knowledge from the set of learned policies to bootstrap its search for new optimal policy. We demonstrate our approach on a simulated mini-golf task where the 7 degrees of freedom Barrett WAM robot arm learns to sequentially putt on different holes in accordance with the playing strategies of the expert.
引用
收藏
页码:3244 / 3250
页数:7
相关论文
共 50 条
  • [21] Survey on Inverse Reinforcement Learning
    Zhang L.-H.
    Liu Q.
    Huang Z.-G.
    Zhu F.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (10): : 4772 - 4803
  • [22] A survey of inverse reinforcement learning
    Stephen Adams
    Tyler Cody
    Peter A. Beling
    Artificial Intelligence Review, 2022, 55 : 4307 - 4346
  • [23] Mapping Language to Programs using Multiple Reward Components with Inverse Reinforcement Learning
    Ghosh, Sayan
    Srivastava, Shashank
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 1449 - 1462
  • [24] Learning Behavior Styles with Inverse Reinforcement Learning
    Lee, Seong Jae
    popovic, Zoran
    ACM TRANSACTIONS ON GRAPHICS, 2010, 29 (04):
  • [25] Reinforcement Learning and Inverse Reinforcement Learning with System 1 and System 2
    Peysakhovich, Alexander
    AIES '19: PROCEEDINGS OF THE 2019 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2019, : 409 - 415
  • [26] Teaching Multiple Inverse Reinforcement Learners
    Melo, Francisco S.
    Lopes, Manuel
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2021, 4
  • [27] Inverse Reinforcement Learning for Text Summarization
    Fu, Yu
    Xiong, Deyi
    Dong, Yue
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 6559 - 6570
  • [28] Reward Identification in Inverse Reinforcement Learning
    Kim, Kuno
    Garg, Shivam
    Shiragur, Kirankumar
    Ermon, Stefano
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [29] Compatible Reward Inverse Reinforcement Learning
    Metelli, Alberto Maria
    Pirotta, Matteo
    Restelli, Marcello
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [30] Training parsers by inverse reinforcement learning
    Gergely Neu
    Csaba Szepesvári
    Machine Learning, 2009, 77 : 303 - 337