DeMoCap: Low-Cost Marker-Based Motion Capture

被引:28
作者
Chatzitofis, Anargyros [1 ,2 ]
Zarpalas, Dimitrios [2 ]
Daras, Petros [2 ]
Kollias, Stefanos [1 ]
机构
[1] Natl Tech Univ Athens, Zografou Campus 9, Athens 15780, Greece
[2] Ctr Res & Technol Hellas, 6th Km Charilaou Thermi, Thessaloniki 57001, Greece
关键词
Motion capture; Low-cost; Marker-based; Depth-based; Pose regression; Multi-view;
D O I
10.1007/s11263-021-01526-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Optical marker-based motion capture (MoCap) remains the predominant way to acquire high-fidelity articulated body motions. We introduce DeMoCap, the first data-driven approach for end-to-end marker-based MoCap, using only a sparse setup of spatio-temporally aligned, consumer-grade infrared-depth cameras. Trading off some of their typical features, our approach is the sole robust option for far lower-cost marker-based MoCap than high-end solutions. We introduce an end-to-end differentiable markers-to-pose model to solve a set of challenges such as under-constrained position estimates, noisy input data and spatial configuration invariance. We simultaneously handle depth and marker detection noise, label and localize the markers, and estimate the 3D pose by introducing a novel spatial 3D coordinate regression technique under a multi-view rendering and supervision concept. DeMoCap is driven by a special dataset captured with 4 spatio-temporally aligned low-cost Intel RealSense D415 sensors and a 24 MXT40S camera professional MoCap system, used as input and ground truth, respectively.
引用
收藏
页码:3338 / 3366
页数:29
相关论文
共 64 条
[21]  
He, P IEEE C COMP VIS PA, P770, DOI DOI 10.1109/CVPR.2016.90
[22]  
He KM, 2020, IEEE T PATTERN ANAL, V42, P386, DOI [10.1109/TPAMI.2018.2844175, 10.1109/ICCV.2017.322]
[23]   Robust Solving of Optical Motion Capture Data by Denoising [J].
Holden, Daniel .
ACM TRANSACTIONS ON GRAPHICS, 2018, 37 (04)
[24]   Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments [J].
Ionescu, Catalin ;
Papava, Dragos ;
Olaru, Vlad ;
Sminchisescu, Cristian .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2014, 36 (07) :1325-1339
[25]   Learnable Triangulation of Human Pose [J].
Iskakov, Karim ;
Burkov, Egor ;
Lempitsky, Victor ;
Malkov, Yury .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :7717-7726
[26]   Total Capture: A 3D Deformation Model for Tracking Faces, Hands, and Bodies [J].
Joo, Hanbyul ;
Simon, Tomas ;
Sheikh, Yaser .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8320-8329
[27]   Intel® RealSense™ Stereoscopic Depth Cameras [J].
Keselman, Leonid ;
Woodfill, John Iselin ;
Grunnet-Jepsen, Anders ;
Bhowmik, Achintya .
2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, :1267-1276
[28]  
Kingma DP, 2014, ADV NEUR IN, V27
[29]   Maximum-Margin Structured Learning with Deep Networks for 3D Human Pose Estimation [J].
Li, Sijin ;
Zhang, Weichen ;
Chan, Antoni B. .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2848-2856
[30]   Microsoft COCO: Common Objects in Context [J].
Lin, Tsung-Yi ;
Maire, Michael ;
Belongie, Serge ;
Hays, James ;
Perona, Pietro ;
Ramanan, Deva ;
Dollar, Piotr ;
Zitnick, C. Lawrence .
COMPUTER VISION - ECCV 2014, PT V, 2014, 8693 :740-755