Human action recognition by leaning pose dictionary

被引:0
作者
Cai, Jiaxin [1 ,2 ]
Feng, Guocan [1 ,2 ]
Tang, Xin [1 ,2 ]
Luo, Zhihong [3 ]
机构
[1] School of Mathematics and Computing Science, Sun Yat-Sen University, Guangzhou, 510275, Guangdong
[2] Guangdong Province Key Laboratory of Computational Science, Guangzhou, 510275, Guangdong
[3] School of Informational Science and Technology, Sun Yat-Sen University, Guangzhou, 510275, Guangdong
来源
Cai, Jiaxin | 1600年 / Chinese Optical Society卷 / 34期
关键词
Behavior recognition; Dictionary learning; Image processing; Local preserving projection; Procrustes shape analysis; Sparse representation;
D O I
10.3788/AOS201434.1215002
中图分类号
学科分类号
摘要
A framework for human action recognition by learning pose dictionary based on human contour representation is proposed. A new pose feature based on Procrustes analysis and local preserving projection is proposed, which can extract shape information from human motion video which is invariant to translation, scaling and rotation. Moreover, it can extract discriminative subspace information when preserving local manifold structure of human pose. After the pose feature is extracted, a human action recognition framework based on pose dictionary learning is proposed. Class-specific dictionaries are trained individually on all training frames of each class to build the whole pose dictionary by concatenating all class-specific dictionaries. The test video is classified with the minimum reconstruction error on the learned dictionary. Experimental results on Weizmann and MuHAVi-MAS14 dataset demonstrate proposed method outperforms most classical methods. Especially, classification rate of this method on MuHAVi-MAS14 dataset achieves a considerable boost compared with that of state-of-the-art approaches. ©, 2014, Chinese Optical Society. All right reserved.
引用
收藏
页数:12
相关论文
共 36 条
  • [1] Bobick A.F., Davis J.W., The recognition of human movement using temporal templates , IEEE Trans Pattern Anal Mach Intell, 23, 3, pp. 257-267, (2001)
  • [2] Xiang T., Gong S., Beyond tracking: modelling activity and understanding behaviour , International Journal of Computer Vision, 67, 1, pp. 21-51, (2006)
  • [3] Chandrashekhar V., Venkatesh K., Action energy images for reliable human action recognition , Proc of Asian Symp on Information Display, pp. 484-487, (2006)
  • [4] Wang L., Suter D., Informative shape representations for human ction recognition , 18th International Conference on Pattern Recognition, pp. 1266-1269, (2006)
  • [5] Weinland D., Ronfard R., Boyer E., Free viewpoint action recognition using motion history volumes , Computer Vision and Image Understanding, 104, 2-3, pp. 249-257, (2007)
  • [6] Cai J.X., Feng G.C., Tang X., Human action recognition using oriented holistic feature , 20th IEEE International Conference on Image Processing, pp. 2420-2424, (2013)
  • [7] Niebles C., Wang H.C., Li F.F., Unsupervised learning of human action categories using spatial-temporal words , International Journal of Computer Vision, 79, 3, pp. 299-318, (2008)
  • [8] Liu J.G., Ali S., Shah M., Recognizing human actions using multiple features , Computer Vision and Pattern Recognition, pp. 1-8, (2008)
  • [9] Alexander K., Marcin M., Cordelia S., A spatio-temporal descriptor based on 3D-gradients , British Machine Vision Conference, pp. 995-1004, (2008)
  • [10] Scovanner P., Saad A., Mubarak S., A 3-dimensional sift descriptor and its application to action recognition , Proceedings of the 15th international conference on Multimedia, pp. 357-360, (2007)