Learning Agile Skills via Adversarial Imitation of Rough Partial Demonstrations

被引:0
|
作者
Li, Chenhao [1 ,2 ]
Vlastelica, Marin [1 ]
Blaes, Sebastian [1 ]
Frey, Jonas [1 ,2 ]
Grimminger, Felix [1 ]
Martius, Georg [1 ]
机构
[1] Max Planck Inst Intelligent Syst, Stuttgart, Germany
[2] Swiss Fed Inst Technol, Robot Syst Lab, Zurich, Switzerland
来源
CONFERENCE ON ROBOT LEARNING, VOL 205 | 2022年 / 205卷
关键词
Adversarial; Imitation Learning; Legged Robots; LOCOMOTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning agile skills is one of the main challenges in robotics. To this end, reinforcement learning approaches have achieved impressive results. These methods require explicit task information in terms of a reward function or an expert that can be queried in simulation to provide a target control output, which limits their applicability. In this work, we propose a generative adversarial method for inferring reward functions from partial and potentially physically incompatible demonstrations for successful skill acquirement where reference or expert demonstrations are not easily accessible. Moreover, we show that by using a Wasserstein GAN formulation and transitions from demonstrations with rough and partial information as input, we are able to extract policies that are robust and capable of imitating demonstrated behaviors. Finally, the obtained skills such as a backflip are tested on an agile quadruped robot called Solo 8 and present faithful replication of hand-held human demonstrations.
引用
收藏
页码:342 / 352
页数:11
相关论文
共 50 条
  • [31] Joint Entity and Event Extraction with Generative Adversarial Imitation Learning
    Zhang, Tongtao
    Ji, Heng
    Sil, Avirup
    DATA INTELLIGENCE, 2019, 1 (02) : 99 - 120
  • [32] Improve generated adversarial imitation learning with reward variance regularization
    Yi-Feng Zhang
    Fan-Ming Luo
    Yang Yu
    Machine Learning, 2022, 111 : 977 - 995
  • [33] Imitation learning from imperfect demonstrations for AUV path tracking and obstacle avoidance
    Chen, Tianhao
    Zhang, Zheng
    Fang, Zheng
    Jiang, Dong
    Li, Guangliang
    OCEAN ENGINEERING, 2024, 298
  • [34] Learning Task-Parameterized Skills From Few Demonstrations
    Zhu, Jihong
    Gienger, Michael
    Kober, Jens
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 4063 - 4070
  • [35] Best-in-class imitation: Non-negative positive-unlabeled imitation learning from imperfect demonstrations
    Zhang, Lin
    Zhu, Fei
    Ling, Xinghong
    Liu, Quan
    INFORMATION SCIENCES, 2022, 601 : 71 - 89
  • [36] Learning by Watching via Keypoint Extraction and Imitation Learning
    Sun, Yin-Tung Albert
    Lin, Hsin-Chang
    Wu, Po-Yen
    Huang, Jung-Tang
    MACHINES, 2022, 10 (11)
  • [37] The Art of Imitation: Learning Long-Horizon Manipulation Tasks From Few Demonstrations
    von Hartz, Jan Ole
    Welschehold, Tim
    Valada, Abhinav
    Boedecker, Joschka
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (12): : 11369 - 11376
  • [38] An imitation learning framework for generating multi-modal trajectories from unstructured demonstrations
    Peng, Jian-Wei
    Hu, Min-Chun
    Chu, Wei-Ta
    NEUROCOMPUTING, 2022, 500 : 712 - 723
  • [39] Urban Vehicle Trajectory Generation Based on Generative Adversarial Imitation Learning
    Wang, Min
    Cui, Jianqun
    Wong, Yew Wee
    Chang, Yanan
    Wu, Libing
    Jin, Jiong
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (12) : 18237 - 18249
  • [40] Developing multi-agent adversarial environment using reinforcement learning and imitation learning
    Ziyao Han
    Yupeng Liang
    Kazuhiro Ohkura
    Artificial Life and Robotics, 2023, 28 : 703 - 709