DeepMoCap: Deep Optical Motion Capture Using Multiple Depth Sensors and Retro-Reflectors

被引:15
作者
Chatzitofis, Anargyros [1 ,2 ]
Zarpalas, Dimitrios [1 ]
Kollias, Stefanos [2 ,3 ]
Daras, Petros [1 ]
机构
[1] Informat Technol Inst, Ctr Res & Technol Hellas, 6th Km Charilaou Thermi, Thessaloniki 57001, Greece
[2] Natl Tech Univ Athens, Sch Elect & Comp Engn, Zografou Campus,Iroon Polytechniou 9, Athens 15780, Greece
[3] Univ Lincoln, Sch Comp Sci, Brayford LN67TS, England
基金
欧盟地平线“2020”;
关键词
motion capture; deep learning; retro-reflectors; retro-reflective markers; multiple depth sensors; low-cost; deep mocap; depth data; 3D data; 3D vision; optical mocap; marker-based mocap; POSE ESTIMATION; RECONSTRUCTION; ACCURACY; TRACKING; KINECT;
D O I
10.3390/s19020282
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
In this paper, a marker-based, single-person optical motion capture method (DeepMoCap) is proposed using multiple spatio-temporally aligned infrared-depth sensors and retro-reflective straps and patches (reflectors). DeepMoCap explores motion capture by automatically localizing and labeling reflectors on depth images and, subsequently, on 3D space. Introducing a non-parametric representation to encode the temporal correlation among pairs of colorized depthmaps and 3D optical flow frames, a multi-stage Fully Convolutional Network (FCN) architecture is proposed to jointly learn reflector locations and their temporal dependency among sequential frames. The extracted reflector 2D locations are spatially mapped in 3D space, resulting in robust 3D optical data extraction. The subject's motion is efficiently captured by applying a template-based fitting technique on the extracted optical data. Two datasets have been created and made publicly available for evaluation purposes; one comprising multi-view depth and 3D optical flow annotated images (DMC2.5D), and a second, consisting of spatio-temporally aligned multi-view depth images along with skeleton, inertial and ground truth MoCap data (DMC3D). The FCN model outperforms its competitors on the DMC2.5D dataset using 2D Percentage of Correct Keypoints (PCK) metric, while the motion capture outcome is evaluated against RGB-D and inertial data fusion approaches on DMC3D, outperforming the next best method by <mml:semantics>4.5%</mml:semantics> in total 3D PCK accuracy.
引用
收藏
页数:26
相关论文
共 52 条
  • [1] An Integrated Platform for Live 3D Human Reconstruction and Motion Capturing
    Alexiadis, Dimitrios S.
    Chatzitofis, Anargyros
    Zioulis, Nikolaos
    Zoidi, Olga
    Louizis, Georgios
    Zarpalas, Dimitrios
    Daras, Petros
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2017, 27 (04) : 798 - 813
  • [2] [Anonymous], 2017, ARXIV171006235
  • [3] [Anonymous], 2017, ARXIV171206316
  • [4] Asteriadis S., 2013, P 6 INT C COMPUTER V, P3, DOI [DOI 10.1145/2466715.2466727, 10.1145/2466715.2466727]
  • [5] Barnes F.P., 1993, U.S. Patent, Patent No. [5,249,106, 524910628]
  • [6] Bodenheimer B., 1997, The process of motion capture: Dealing with the data Computer Animation and Simulation'97, P3
  • [7] Temporal-spatial reach parameters derived from inertial sensors: Comparison to 3D marker-based motion capture
    Cahill-Rowley, Katelyn
    Rose, Jessica
    [J]. JOURNAL OF BIOMECHANICS, 2017, 52 : 11 - 16
  • [8] Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields
    Cao, Zhe
    Simon, Tomas
    Wei, Shih-En
    Sheikh, Yaser
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 1302 - 1310
  • [9] Chen X., 2000, Camera placement considering occlusion for robust motion capture, V2, P2
  • [10] Kinect as a Tool for Gait Analysis: Validation of a Real-Time Joint Extraction AlgorithmWorking in Side View
    Cippitelli, Enea
    Gasparrini, Samuele
    Spinsante, Susanna
    Gambi, Ennio
    [J]. SENSORS, 2015, 15 (01) : 1417 - 1434