Effective human action recognition using global and local offsets of skeleton joints

被引:13
作者
Sun, Bin [1 ]
Kong, Dehui [1 ]
Wang, Shaofan [1 ]
Wang, Lichun [1 ]
Wang, Yuping [1 ]
Yin, Baocai [2 ]
机构
[1] Beijing Univ Technol, BJUT Fac Informat Technol, Beijing Adv Innovat Ctr Future Internet Technol, Beijing Key Lab Multimedia & Intelligent Software, Beijing 100124, Peoples R China
[2] Dalian Univ Technol, Coll Comp Sci & Technol, Fac Elect Informat & Elect Engn, Dalian 116024, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Action recognition; Skeleton joints; Offsets; Histogram representation; Naive-Bayes-Nearest-Neighbor; ENSEMBLE; FEATURES; SPACE;
D O I
10.1007/s11042-018-6370-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Human action recognition based on 3D skeleton joints is an important yet challenging task. While many research work are devoted to 3D action recognition, they mainly suffer from two problems: complex model representation and low implementation efficiency. To tackle these problems, we propose an effective and efficient framework for 3D action recognition using a global-and-local histogram representation model. Our method consists of a global-and-local featuring phase, a saturation based histogram representation phase, and a classification phase. The global-and-local featuring phase captures the global feature and local feature of each action sequence using the joint displacement between the current frame and the first frame, and the joint displacement between pairwise fixed-skip frames, respectively. The saturation based histogram representation phase captures the histogram representation of each joint considering the motion independence of joints and saturation of each histogram bin. The classification phase measures the distance of each joint histogram-to-class. Besides, we produce a novel action dataset called BJUT Kinect dataset, which consists of multi-period motion clips and intra-class variations. We compare our method with many state-of-the-art methods on BJUT Kinect dataset, UCF Kinect dataset, Florence 3D action dataset, MSR-Action3D dataset, and NTU RGB+D Dataset. The results show that our method achieves both higher accuracy and efficiency for 3D action recognition.
引用
收藏
页码:6329 / 6353
页数:25
相关论文
共 58 条
[41]  
Sung JY, 2012, IEEE INT CONF ROBOT, P842, DOI 10.1109/ICRA.2012.6224591
[42]   Differential Recurrent Neural Networks for Action Recognition [J].
Veeriah, Vivek ;
Zhuang, Naifan ;
Qi, Guo-Jun .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :4041-4049
[43]   Human Action Recognition by Representing 3D Skeletons as Points in a Lie Group [J].
Vemulapalli, Raviteja ;
Arrate, Felipe ;
Chellappa, Rama .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :588-595
[44]   On the improvement of human action recognition from depth map sequences using Space-Time Occupancy Patterns [J].
Vieira, Antonio W. ;
Nascimento, Erickson R. ;
Oliveira, Gabriel L. ;
Liu, Zicheng ;
Campos, Mario F. M. .
PATTERN RECOGNITION LETTERS, 2014, 36 :221-227
[45]  
Vieira AW, 2012, PROGR PATTERN RECOGN, P252, DOI [DOI 10.1007/978-3-642-33275-3, DOI 10.1007/978-3-642-33275-331, DOI 10.1007/978-3-642-33275]
[46]   Learning Actionlet Ensemble for 3D Human Action Recognition [J].
Wang, Jiang ;
Liu, Zicheng ;
Wu, Ying ;
Yuan, Junsong .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2014, 36 (05) :914-927
[47]  
Wang J, 2012, LECT NOTES COMPUT SC, V7573, P872, DOI 10.1007/978-3-642-33709-3_62
[48]   Mining Actionlet Ensemble for Action Recognition with Depth Cameras [J].
Wang, Jiang ;
Liu, Zicheng ;
Wu, Ying ;
Yuan, Junsong .
2012 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2012, :1290-1297
[49]   Graph Based Skeleton Motion Representation and Similarity Measurement for Action Recognition [J].
Wang, Pei ;
Yuan, Chunfeng ;
Hu, Weiming ;
Li, Bing ;
Zhang, Yanning .
COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 :370-385
[50]  
Xia Lu, 2012, IEEE COMP SOC C COMP, P20, DOI [10.1109/CVPRW.2012.6239233, DOI 10.1109/CVPRW.2012.6239233]