Exploiting temporal context for 3D human pose estimation in the wild

被引:178
作者
Arnab, Anurag [1 ,2 ]
Doersch, Carl [2 ]
Zisserman, Andrew [1 ,2 ]
机构
[1] Univ Oxford, Oxford, England
[2] DeepMind, London, England
来源
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019) | 2019年
关键词
RECONSTRUCTION; SHAPE;
D O I
10.1109/CVPR.2019.00351
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a bundle-adjustment-based algorithm for recovering accurate 3D human pose and meshes from monocular videos. Unlike previous algorithms which operate on single frames, we show that reconstructing a person over an entire sequence gives extra constraints that can resolve ambiguities. This is because videos often give multiple views of a person, yet the overall body shape does not change and 3D positions vary slowly. Our method improves not only on standard mocap-based datasets like Human 3.6M - where we show quantitative improvements - but also on challenging in-the-wild datasets such as Kinetics. Building upon our algorithm, we present a new dataset of more than 3 million frames of YouTube videos from Kinetics with automatically generated 3D poses and meshes. We show that retraining a single frame 3D pose estimator on this data improves accuracy on both real-world and mocap data by evaluating on the 3DPW and HumanEVA datasets.
引用
收藏
页码:3390 / 3399
页数:10
相关论文
共 57 条
[1]  
Akhter I., 2015, CVPR
[2]  
Andriluka M., 2014, CVPR
[3]   Monocular 3D Pose Estimation and Tracking by Detection [J].
Andriluka, Mykhaylo ;
Roth, Stefan ;
Schiele, Bernt .
2010 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2010, :623-630
[4]   SCAPE: Shape Completion and Animation of People [J].
Anguelov, D ;
Srinivasan, P ;
Koller, D ;
Thrun, S ;
Rodgers, J ;
Davis, J .
ACM TRANSACTIONS ON GRAPHICS, 2005, 24 (03) :408-416
[5]  
[Anonymous], 2008, PROC INT C NEURAL IN
[6]  
[Anonymous], 2018, CVPR
[7]  
[Anonymous], 2016, 3DV
[8]  
[Anonymous], 2018, ECCV
[9]  
[Anonymous], 2014, Advances in neural information processing systems
[10]  
[Anonymous], 2017, P CVPR