A Model-Based System for Real-Time Articulated Hand Tracking Using a Simple Data Glove and a Depth Camera

被引:13
作者
Jiang, Linjun [1 ,2 ]
Xia, Hailun [1 ,2 ]
Guo, Caili [1 ,2 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Informat & Commun Engn, Beijing Key Lab Network Syst Architecture & Conve, Beijing 100876, Peoples R China
[2] Beijing Lab Adv Informat Networks, Beijing 100876, Peoples R China
基金
中国国家自然科学基金;
关键词
articulated hand tracking; multi-model; data glove; depth camera; model-fitting; real-time; POSE ESTIMATION; REGRESSION; FUSION; KINECT;
D O I
10.3390/s19214680
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Tracking detailed hand motion is a fundamental research topic in the area of human-computer interaction (HCI) and has been widely studied for decades. Existing solutions with single-model inputs either require tedious calibration, are expensive or lack sufficient robustness and accuracy due to occlusions. In this study, we present a real-time system to reconstruct the exact hand motion by iteratively fitting a triangular mesh model to the absolute measurement of hand from a depth camera under the robust restriction of a simple data glove. We redefine and simplify the function of the data glove to lighten its limitations, i.e., tedious calibration, cumbersome equipment, and hampering movement and keep our system lightweight. For accurate hand tracking, we introduce a new set of degrees of freedom (DoFs), a shape adjustment term for personalizing the triangular mesh model, and an adaptive collision term to prevent self-intersection. For efficiency, we extract a strong pose-space prior to the data glove to narrow the pose searching space. We also present a simplified approach for computing tracking correspondences without the loss of accuracy to reduce computation cost. Quantitative experiments show the comparable or increased accuracy of our system over the state-of-the-art with about 40% improvement in robustness. Besides, our system runs independent of Graphic Processing Unit (GPU) and reaches 40 frames per second (FPS) at about 25% Central Processing Unit (CPU) usage.
引用
收藏
页数:29
相关论文
共 46 条
[1]  
[Anonymous], 2013, P GRAPHICS INTERFACE
[2]  
[Anonymous], 2015, P BRIT MACHINE VISIO
[3]  
[Anonymous], 2017, ARXIV170402224
[4]   Robust Hand Motion Tracking through Data Fusion of 5DT Data Glove and Nimble VR Kinect Camera Measurements [J].
Arkenbout, Ewout A. ;
de Winter, Joost C. F. ;
Breedveld, Paul .
SENSORS, 2015, 15 (12) :31644-31671
[5]  
Ballan L., 2014, P EUR C COMP VIS FIR, P640
[6]   SHPR-Net: Deep Semantic Hand Pose Regression From Point Clouds [J].
Chen, Xinghao ;
Wang, Guijin ;
Zhang, Cairong ;
Kim, Tae-Kyun ;
Ji, Xiangyang .
IEEE ACCESS, 2018, 6 :43425-43439
[7]   A survey of glove-based systems and their applications [J].
Dipietro, Laura ;
Sabatini, Angelo M. ;
Dario, Paolo .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (04) :461-482
[8]   CrosslnfoNet: Multi-Task Information Sharing Based Hand Pose Estimation [J].
Du, Kuo ;
Lin, Xiangbo ;
Sun, Yi ;
Ma, Xiaohong .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9888-9897
[9]  
Fleishman Shachar, 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), P28, DOI 10.1109/CVPRW.2015.7301345
[10]  
Ganapathi V, 2012, LECT NOTES COMPUT SC, V7577, P738, DOI 10.1007/978-3-642-33783-3_53