Pose Guided Human Video Generation

被引:28
作者
Yang, Ceyuan [1 ]
Wang, Zhe [2 ]
Zhu, Xinge [1 ]
Huang, Chen [3 ]
Shi, Jianping [2 ]
Lin, Dahua [1 ]
机构
[1] CUHK, CUHK SenseTime Joint Lab, Hong Kong, Peoples R China
[2] SenseTime Res, Beijing, Peoples R China
[3] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
来源
COMPUTER VISION - ECCV 2018, PT X | 2018年 / 11214卷
关键词
Human video generation; Pose synthesis; Generation adversarial network;
D O I
10.1007/978-3-030-01249-6_13
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the emergence of Generative Adversarial Networks, video synthesis has witnessed exceptional breakthroughs. However, existing methods lack a proper representation to explicitly control the dynamics in videos. Human pose, on the other hand, can represent motion patterns intrinsically and interpretably, and impose the geometric constraints regardless of appearance. In this paper, we propose a pose guided method to synthesize human videos in a disentangled way: plausible motion prediction and coherent appearance generation. In the first stage, a Pose Sequence Generative Adversarial Network (PSGAN) learns in an adversarial manner to yield pose sequences conditioned on the class label. In the second stage, a Semantic Consistent Generative Adversarial Network (SCGAN) generates video frames from the poses while preserving coherent appearances in the input image. By enforcing semantic consistency between the generated and ground-truth poses at a high feature level, our SCGAN is robust to noisy or abnormal poses. Extensive experiments on both human action and human face datasets manifest the superiority of the proposed method over other state-of-the-arts.
引用
收藏
页码:204 / 219
页数:16
相关论文
共 40 条
[1]  
[Anonymous], 2017, ICCV
[2]  
[Anonymous], 2018, CVPR
[3]  
[Anonymous], 2016, NIPS
[4]  
[Anonymous], 2014, CVPR
[5]  
[Anonymous], 2018, AAAI
[6]  
[Anonymous], 2015, CVPR
[7]  
[Anonymous], 2016, ARXIV E PRINTS
[8]  
[Anonymous], 2017, NIPS
[9]  
[Anonymous], 2017, ARXIV170107875
[10]  
[Anonymous], 2017, INT C MACH LEARN ICM