Architecture and Parameter Analysis to Convolutional Neural Network for Hand Tracking

被引:0
作者
Zhou, Hui [1 ,2 ]
Yang, Minghao [1 ,2 ,4 ]
Pan, Hang [1 ,2 ]
Tang, Renjun [1 ,2 ]
Qiang, Baohua [1 ]
Chen, Jinlong [1 ]
Tao, Jianhua [2 ,3 ,4 ]
机构
[1] Guilin Univ Elect Technol, Guangxi Key Lab Cryptog & Informat Secur, Guilin 541004, Guangxi, Peoples R China
[2] Chinese Acad Sci, Inst Automat, Natl Key Lab Pattern Recognit, Beijing 100190, Peoples R China
[3] Univ Chinese Acad Sci, Beijing 100190, Peoples R China
[4] Chinese Acad Sci, Inst Automat, Res Ctr Brain Inspired Intelligence, Beijing 100190, Peoples R China
来源
CLOUD COMPUTING AND SECURITY, PT VI | 2018年 / 11068卷
关键词
Hand tracking; Convolutional neural network; Deep learning; Data augmentation; Dropout;
D O I
10.1007/978-3-030-00021-9_39
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Currently, the hand tracking based on deep learning has made good progress, but these literatures have less influence on the tracking accuracy of Convolutional Neural Network (CNN) architecture and parameters. In this paper, we proposed a new method to analyze the influence factors of gesture tracking. Firstly, we establish the gesture image and corresponding gesture parameter database based on virtual 3D human hand, on which the convolutional neural network models are constructed, after that we research some related factors, such as network structure, iteration times, data augmentation and Dropout, etc., that affect the performance of hand tracking. Finally we evaluate the objective parameters of the virtual hand, and make the subjective evaluation of the real hand extracted in the video. The results show that, on the premise of the fixed training amount of the hand, the effect of increasing the number of convolutional cores or convolution layers on the accuracy of the real gesture is not obvious, the data augmentation is obvious. For the real gesture, when the number of iterations and the Dropout ratio is about 20%-30%, good results can be obtained. This work provides the foundation for future application research on hand tracking.
引用
收藏
页码:429 / 439
页数:11
相关论文
共 14 条
[1]  
[Anonymous], 2008, P 2008 COMP VIS PATT
[2]  
[Anonymous], 2014, ACM T GRAPHIC, DOI DOI 10.1145/2629500
[3]  
Bouthillier X., 2016, P INT C LEARNING REP, P1
[4]   A survey of glove-based systems and their applications [J].
Dipietro, Laura ;
Sabatini, Angelo M. ;
Dario, Paolo .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (04) :461-482
[5]   Robust 3D Hand Pose Estimation in Single Depth Images: from Single-View CNN to Multi-View CNNs [J].
Ge, Liuhao ;
Liang, Hui ;
Yuan, Junsong ;
Thalmann, Daniel .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3593-3601
[6]  
Krizhevsky A., 2017, COMMUN ACM, V60, P84, DOI [DOI 10.1145/3065386, 10.1145/3065386]
[7]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324
[8]  
Navaratnam R., 2007, P INT C COMPUTER VIS, P1
[9]   Realtime and Robust Hand Tracking from Depth [J].
Qian, Chen ;
Sun, Xiao ;
Wei, Yichen ;
Tang, Xiaoou ;
Sun, Jian .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :1106-1113
[10]   Accurate, Robust, and Flexible Real-time Hand Tracking [J].
Sharp, Toby ;
Keskin, Cem ;
Robertson, Duncan ;
Taylor, Jonathan ;
Shotton, Jamie ;
Kim, David ;
Rhemann, Christoph ;
Leichter, Ido ;
Vinnikov, Alon ;
Wei, Yichen ;
Freedman, Daniel ;
Kohli, Pushmeet ;
Krupka, Eyal ;
Fitzgibbon, Andrew ;
Izadi, Shahram .
CHI 2015: PROCEEDINGS OF THE 33RD ANNUAL CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2015, :3633-3642