Transfer in Inverse Reinforcement Learning for Multiple Strategies

被引:0
|
作者
Tanwani, Ajay Kumar [1 ]
Billard, Aude [1 ]
机构
[1] Ecole Polytech Fed Lausanne, LASA, CH-1015 Lausanne, Switzerland
来源
2013 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS) | 2013年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the problem of incrementally learning different strategies of performing a complex sequential task from multiple demonstrations of an expert or a set of experts. While the task is the same, each expert differs in his/her way of performing it. We assume that this variety across experts' demonstration is due to the fact that each expert/strategy is driven by a different reward function, where reward function is expressed as a linear combination of a set of known features. Consequently, we can learn all the expert strategies by forming a convex set of optimal deterministic policies, from which one can match any unseen expert strategy drawn from this set. Instead of learning from scratch every optimal policy in this set, the learner transfers knowledge from the set of learned policies to bootstrap its search for new optimal policy. We demonstrate our approach on a simulated mini-golf task where the 7 degrees of freedom Barrett WAM robot arm learns to sequentially putt on different holes in accordance with the playing strategies of the expert.
引用
收藏
页码:3244 / 3250
页数:7
相关论文
共 50 条
  • [41] Hierarchical Bayesian Inverse Reinforcement Learning
    Choi, Jaedeug
    Kim, Kee-Eung
    IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (04) : 793 - 805
  • [42] Inverse reinforcement learning in contextual MDPs
    Stav Belogolovsky
    Philip Korsunsky
    Shie Mannor
    Chen Tessler
    Tom Zahavy
    Machine Learning, 2021, 110 : 2295 - 2334
  • [43] Inverse Reinforcement Learning from Failure
    Shiarlis, Kyriacos
    Messias, Joao
    Whiteson, Shimon
    AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 1060 - 1068
  • [44] Theory of mind as inverse reinforcement learning
    Jara-Ettinger, Julian
    CURRENT OPINION IN BEHAVIORAL SCIENCES, 2019, 29 : 105 - 110
  • [45] Receding Horizon Inverse Reinforcement Learning
    Xu, Yiqing
    Gao, Wei
    Hsu, David
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [46] An ensemble method for inverse reinforcement learning
    Lin, Jin-Ling
    Hwang, Kao-Shing
    Shi, Haobin
    Pan, Wei
    INFORMATION SCIENCES, 2020, 512 (512) : 518 - 532
  • [47] Inverse Reinforcement Learning in Swarm Systems
    Sosic, Adrian
    KhudaBukhsh, Wasiur R.
    Zoubir, Abdelhak M.
    Koeppl, Heinz
    AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 1413 - 1421
  • [48] Inverse Reinforcement Learning in Relational Domains
    Munzer, Thibaut
    Piot, Bilal
    Geist, Matthieu
    Pietquin, Olivier
    Lopes, Manuel
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 3735 - 3741
  • [49] Hierarchical Adversarial Inverse Reinforcement Learning
    Chen, Jiayu
    Lan, Tian
    Aggarwal, Vaneet
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17549 - 17558
  • [50] Curricular Subgoals for Inverse Reinforcement Learning
    Liu, Shunyu
    Qing, Yunpeng
    Xu, Shuqi
    Wu, Hongyan
    Zhang, Jiangtao
    Cong, Jingyuan
    Chen, Tianhao
    Liu, Yun-Fu
    Song, Mingli
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2025, 26 (03) : 3016 - 3027