Deep Adversarial Imitation Learning of Locomotion Skills from One-shot Video Demonstration

被引:0
作者
Zhang, Huiwen [1 ,2 ,3 ,4 ]
Liu, Yuwang [1 ,2 ,3 ]
Zhou, Weijia [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Shenyang Inst Automat, Shenyang, Peoples R China
[2] Chinese Acad Sci, Inst Robot, Shenyang, Peoples R China
[3] Chinese Acad Sci, Inst Intelligent Mfg, Shenyang, Peoples R China
[4] Univ Chinese Acad Sci, Shenyang, Peoples R China
来源
2019 9TH IEEE ANNUAL INTERNATIONAL CONFERENCE ON CYBER TECHNOLOGY IN AUTOMATION, CONTROL, AND INTELLIGENT SYSTEMS (IEEE-CYBER 2019) | 2019年
基金
中国国家自然科学基金;
关键词
imitation learning; GAN; pose estimation; locomotion skills;
D O I
10.1109/cyber46603.2019.9066512
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Traditional imitation learning approaches usually collect demonstrations by teleoperation, kinesthetic teaching or precisely calibrated motion capture devices. These teaching interfaces are cumbersome and subject to the constraints of the environment and robot structures. Learning from observation adopts the idea that the robot can learn skills by observing human's behaviors, which is more convenient and preferable. However, learning from observation shows great challenges since it involves understanding of the environment and human actions, as well as solving the retarget problem. This paper presents a way to learn locomotion skills from a single video demonstration. We first leverage a weak supervised method to extract the pose feature from the experts, and then learn a joint position controller trying to match this feature by using the general adversarial network (GAN). This approach avoids cumbersome demonstrations, and more importantly, GAN can generalize learned skills to different subjects. We evaluated our method on a walking task executed by a 56 -degree-of-freedom (DOE) humanoid robot. The experiment demonstrate that the vision -based imitation learning algorithm can be applied to high -dimensional robot task and achieve comparable performance to methods by using finely calibrated motion capture data, which are of great significance for the research on human -robot interaction and robot skill acquisition.
引用
收藏
页码:1257 / 1261
页数:5
相关论文
共 24 条
[1]   A survey of robot learning from demonstration [J].
Argall, Brenna D. ;
Chernova, Sonia ;
Veloso, Manuela ;
Browning, Brett .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2009, 57 (05) :469-483
[2]  
Behbahani F., 2018, ARXIV181103516CSSTAT
[3]  
Dhariwal Prafulla, 2017, OpenAI baselines
[4]  
Ebert F., 2017, CoRL
[5]  
Finn Chelsea, 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA), P2786, DOI 10.1109/ICRA.2017.7989324
[6]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[7]   Dynamical System Modulation for Robot Learning via Kinesthetic Demonstrations [J].
Hersch, Micha ;
Guenter, Florent ;
Calinon, Sylvain ;
Billard, Aude .
IEEE TRANSACTIONS ON ROBOTICS, 2008, 24 (06) :1463-1467
[8]  
Ho J, 2016, Adv. Neural Inf. Process. Syst., V29, P4565
[9]  
Kalashnikov D., 2018, C ROB LEARN, P651
[10]   End-to-end Recovery of Human Shape and Pose [J].
Kanazawa, Angjoo ;
Black, Michael J. ;
Jacobs, David W. ;
Malik, Jitendra .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7122-7131