An imitation learning framework for generating multi-modal trajectories from unstructured demonstrations

被引:3
|
作者
Peng, Jian-Wei [1 ]
Hu, Min-Chun [2 ]
Chu, Wei-Ta [1 ]
机构
[1] Natl Cheng Kung Univ, Dept Comp Sci & Informat Engn, Tainan, Taiwan
[2] Natl Tsing Hua Univ, Dept Comp Sci, Hsinchu, Taiwan
关键词
Trajectory generation; Motion synthesis; Imitation learning; Reinforcement learning; Generative adversarial networks; HUMAN MOTION PREDICTION;
D O I
10.1016/j.neucom.2022.05.076
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The main challenge of the trajectory generation problem is to generate long-term as well as diverse tra-jectories. Generative Adversarial Imitation Learning (GAIL) is a well-known model-free imitation learning algorithm that can be utilized to generate trajectory data, while vanilla GAIL would fail to capture multi -modal demonstrations. Recent methods propose latent variable models to solve this problem; however, previous works may have a mode missing problem. In this work, we propose a novel method to generate long-term trajectories that are controllable by a continuous latent variable based on GAIL and a condi-tional Variational Autoencoder (cVAE). We further assume that subsequences of the same trajectory should be encoded to similar locations in the latent space. Therefore, we introduce a contrastive loss in the training of the encoder. In our motion synthesis task, we propose to first construct a low-dimensional motion manifold by using a VAE to reduce the burden of our imitation learning model. Our experimental results show that the proposed model outperforms the state-of-the-art methods and can be applied to motion synthesis.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:712 / 723
页数:12
相关论文
共 41 条
  • [1] BAGAIL: Multi-modal imitation learning from imbalanced demonstrations
    Gu, Sijia
    Zhu, Fei
    NEURAL NETWORKS, 2024, 174
  • [2] Latent Segmentation of Stock Trading Strategies Using Multi-Modal Imitation Learning
    Maeda, Iwao
    deGraw, David
    Kitano, Michiharu
    Matsushima, Hiroyasu
    Izumi, Kiyoshi
    Sakaji, Hiroki
    Kato, Atsuo
    JOURNAL OF RISK AND FINANCIAL MANAGEMENT, 2020, 13 (11)
  • [3] M3IL: Multi-Modal Meta-Imitation Learning
    Zhang X.
    Matsushima T.
    Matsuo Y.
    Iwasawa Y.
    Transactions of the Japanese Society for Artificial Intelligence, 2023, 38 (02)
  • [4] Adversarial Imitation Learning from State-only Demonstrations
    Torabi, Faraz
    Warnell, Garrett
    Stone, Peter
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2229 - 2231
  • [5] Multi-Modal Legged Locomotion Framework With Automated Residual Reinforcement Learning
    Yu, Chen
    Rosendo, Andre
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04) : 10312 - 10319
  • [6] A Novel Robust Imitation Learning Framework for Complex Skills With Limited Demonstrations
    Wang, Weiyong
    Zeng, Chao
    Zhan, Hong
    Yang, Chenguang
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, : 3947 - 3959
  • [7] Model predictive optimization for imitation learning from demonstrations
    Hu, Yingbai
    Cui, Mingyang
    Duan, Jianghua
    Liu, Wenjun
    Huang, Dianye
    Knoll, Alois
    Chen, Guang
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 163
  • [8] Skill Disentanglement for Imitation Learning from Suboptimal Demonstrations
    Zhao, Tianxiang
    Yu, Wenchao
    Wang, Suhang
    Wang, Lu
    Zhang, Xiang
    Chen, Yuncong
    Liu, Yanchi
    Cheng, Wei
    Chen, Haifeng
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 3513 - 3524
  • [9] Learning to Simulate Vehicle Trajectories from Demonstrations
    Zheng, Guanjie
    Liu, Hanyang
    Xu, Kai
    Li, Zhenhui
    2020 IEEE 36TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2020), 2020, : 1822 - 1825
  • [10] MMDP: A Mobile-IoT Based Multi-Modal Reinforcement Learning Service Framework
    Wang, Puming
    Yang, Laurence T.
    Li, Jintao
    Li, Xue
    Zhou, Xiaokang
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2020, 13 (04) : 675 - 684