Option-Aware Adversarial Inverse Reinforcement Learning for Robotic Control

被引:2
|
作者
Chen, Jiayu [1 ]
Lan, Tian [3 ]
Aggarwal, Vaneet [1 ,2 ]
机构
[1] Purdue Univ, Sch Ind Engn, W Lafayette, IN 47907 USA
[2] KAUST, CS Dept, Thuwal, Saudi Arabia
[3] George Washington Univ, Dept Elect & Comp Engn, Washington, DC 20052 USA
关键词
NEURAL-NETWORKS;
D O I
10.1109/ICRA48891.2023.10160374
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Hierarchical Imitation Learning (HIL) has been proposed to recover highly-complex behaviors in long-horizon tasks from expert demonstrations by modeling the task hierarchy with the option framework. Existing methods either overlook the causal relationship between the subtask and its corresponding policy or cannot learn the policy in an end-to-end fashion, which leads to suboptimality. In this work, we develop a novel HIL algorithm based on Adversarial Inverse Reinforcement Learning and adapt it with the Expectation-Maximization algorithm in order to directly recover a hierarchical policy from the unannotated demonstrations. Further, we introduce a directed information term to the objective function to enhance the causality and propose a Variational Autoencoder framework for learning with our objectives in an end-to-end fashion. Theoretical justifications and evaluations on challenging robotic control tasks are provided to show the superiority of our algorithm. The codes are available at https://github.com/LucasCJYSDL/HierAIRL.
引用
收藏
页码:5902 / 5908
页数:7
相关论文
共 50 条
  • [41] From inverse optimal control to inverse reinforcement learning: A historical review
    Ab Azar, Nematollah
    Shahmansoorian, Aref
    Davoudi, Mohsen
    ANNUAL REVIEWS IN CONTROL, 2020, 50 : 119 - 138
  • [42] A Reinforcement Learning Approach to Health Aware Control Strategy
    Jha, Mayank S.
    Weber, Philippe
    Theilliol, Didier
    Ponsart, Jean-Christophe
    Maquin, Didier
    2019 27TH MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION (MED), 2019, : 171 - 176
  • [43] Reinforcement Learning Control for a Robotic Manipulator with Unknown Deadzone
    Li, Yanan
    Xiao, Shengtao
    Ge, Shuzhi Sam
    2014 11TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2014, : 593 - 598
  • [44] A Reinforcement Learning Neural Network for Robotic Manipulator Control
    Hu, Yazhou
    Si, Bailu
    NEURAL COMPUTATION, 2018, 30 (07) : 1983 - 2004
  • [45] Linguistic Lyapunov reinforcement learning control for robotic manipulators
    Kumar, Abhishek
    Sharma, Rajneesh
    NEUROCOMPUTING, 2018, 272 : 84 - 95
  • [46] Reinforcement Learning for Robotic Safe Control with Force Sensing
    Lin, Nan
    Zhang, Linrui
    Chen, Yuxuan
    Zhu, Yujun
    Chen, Ruoxi
    Wu, Peichen
    Chen, Xiaoping
    2019 WORLD ROBOT CONFERENCE SYMPOSIUM ON ADVANCED ROBOTICS AND AUTOMATION (WRC SARA 2019), 2019, : 148 - 153
  • [47] Adversarial Reinforcement Learning in Dynamic Channel Access and Power Control
    Wang, Fen
    Gursoy, M. Cenk
    Velipasalar, Senem
    2021 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2021,
  • [48] Inverse reinforcement learning control for building energy management
    Dey, Sourav
    Marzullo, Thibault
    Henze, Gregor
    ENERGY AND BUILDINGS, 2023, 286
  • [49] Inverse Reinforcement Learning Control for Linear Multiplayer Games
    Lian, Bosen
    Donge, Vrushabh S.
    Lewis, Frank L.
    Chai, Tianyou
    Davoudi, Ali
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 2839 - 2844
  • [50] COST AWARE ADVERSARIAL LEARNING
    De Silva, Shashini
    Kim, Jinsub
    Raich, Raviv
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3587 - 3591