Option-Aware Adversarial Inverse Reinforcement Learning for Robotic Control

被引:2
|
作者
Chen, Jiayu [1 ]
Lan, Tian [3 ]
Aggarwal, Vaneet [1 ,2 ]
机构
[1] Purdue Univ, Sch Ind Engn, W Lafayette, IN 47907 USA
[2] KAUST, CS Dept, Thuwal, Saudi Arabia
[3] George Washington Univ, Dept Elect & Comp Engn, Washington, DC 20052 USA
关键词
NEURAL-NETWORKS;
D O I
10.1109/ICRA48891.2023.10160374
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Hierarchical Imitation Learning (HIL) has been proposed to recover highly-complex behaviors in long-horizon tasks from expert demonstrations by modeling the task hierarchy with the option framework. Existing methods either overlook the causal relationship between the subtask and its corresponding policy or cannot learn the policy in an end-to-end fashion, which leads to suboptimality. In this work, we develop a novel HIL algorithm based on Adversarial Inverse Reinforcement Learning and adapt it with the Expectation-Maximization algorithm in order to directly recover a hierarchical policy from the unannotated demonstrations. Further, we introduce a directed information term to the objective function to enhance the causality and propose a Variational Autoencoder framework for learning with our objectives in an end-to-end fashion. Theoretical justifications and evaluations on challenging robotic control tasks are provided to show the superiority of our algorithm. The codes are available at https://github.com/LucasCJYSDL/HierAIRL.
引用
收藏
页码:5902 / 5908
页数:7
相关论文
共 50 条
  • [1] Adversarial Option-Aware Hierarchical Imitation Learning
    Jing, Mingxuan
    Huang, Wenbing
    Sunk, Fuchun
    Ma, Xiaojian
    Kong, Tao
    Gan, Chuang
    Li, Lei
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [2] Safety-Aware Adversarial Inverse Reinforcement Learning for Highway Autonomous Driving
    Li, Fangjian
    Wagner, John
    Wang, Yue
    Journal of Autonomous Vehicles and Systems, 2021, 1 (04):
  • [3] Multiagent Adversarial Inverse Reinforcement Learning
    Wei, Ermo
    Wicke, Drew
    Luke, Sean
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2265 - 2266
  • [4] Hierarchical Adversarial Inverse Reinforcement Learning
    Chen, Jiayu
    Lan, Tian
    Aggarwal, Vaneet
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17549 - 17558
  • [5] Inverse Reinforcement Learning for Adversarial Apprentice Games
    Lian, Bosen
    Xue, Wenqian
    Lewis, Frank L.
    Chai, Tianyou
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 4596 - 4609
  • [6] Option compatible reward inverse reinforcement learning
    Hwang, Rakhoon
    Lee, Hanjin
    Hwang, Hyung Ju
    PATTERN RECOGNITION LETTERS, 2022, 154 : 83 - 89
  • [7] Learning Aircraft Pilot Skills by Adversarial Inverse Reinforcement Learning
    Suzuki, Kaito
    Uemura, Tsuneharu
    Tsuchiya, Takeshi
    Beppu, Hirofumi
    Hazui, Yusuke
    Ono, Hitoi
    2023 ASIA-PACIFIC INTERNATIONAL SYMPOSIUM ON AEROSPACE TECHNOLOGY, VOL I, APISAT 2023, 2024, 1050 : 1431 - 1441
  • [8] Multi-Agent Adversarial Inverse Reinforcement Learning
    Yu, Lantao
    Song, Jiaming
    Ermon, Stefano
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [9] Underactuated MSV path following control via stable adversarial inverse reinforcement learning
    Li, Lingyu
    Ma, Yong
    Wu, Defeng
    OCEAN ENGINEERING, 2024, 299
  • [10] Inverse-Inverse Reinforcement Learning. How to Hide Strategy from an Adversarial Inverse Reinforcement Learner
    Pattanayak, Kunal
    Krishnamurthy, Vikram
    Berry, Christopher
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 3631 - 3636