A novel hierarchical framework for human action recognition

被引:54
作者
Chen, Hongzhao [1 ]
Wang, Guijin [1 ]
Xue, Jing-Hao [2 ]
He, Li [1 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing 100084, Peoples R China
[2] UCL, Dept Stat Sci, Mortimer St, London WC1E 6BT, England
关键词
Action recognition; 3D skeleton; Hierarchical framework; Part-based; Time scale; Action graphs; POSE ESTIMATION; REPRESENTATION; JOINTS;
D O I
10.1016/j.patcog.2016.01.020
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a novel two-level hierarchical framework for three-dimensional (3D) skeleton based action recognition, in order to tackle the challenges of high intra-class variance, movement speed variability and high computational costs of action recognition. In the first level, a new part-based clustering module is proposed. In this module, we introduce a part-based five-dimensional (5D) feature vector to explore the most relevant joints of body parts in each action sequence, upon which action sequences are automatically clustered and the high intra-class variance is mitigated. In the second level, there are two modules, motion feature extraction and action graphs. In the module of motion feature extraction, we utilize the cluster-relevant joints only and present a new statistical principle to decide the time scale of motion features, to reduce computational costs and adapt to variable movement speed. In the action graphs module, we exploit these 3D skeleton-based motion features to build action graphs, and devise a new score function based on maximum-likelihood estimation for action graph-based recognition. Experiments on the Microsoft Research Action3D dataset and the University of Texas Kinect Action dataset demonstrate that our method is superior or at least comparable to other state-of-the-art methods, achieving 95.56% recognition rate on the former dataset and 95.96% on the latter one. (C) 2016 Elsevier Ltd. All rights reserved.
引用
收藏
页码:148 / 159
页数:12
相关论文
共 25 条
  • [1] [Anonymous], 2013, P 23 INT JOINT C ART
  • [2] [Anonymous], 2012, P ACM INT C MULT NAR, DOI DOI 10.1145/2393347.2396382
  • [3] Ongoing human action recognition with motion capture
    Barnachon, Mathieu
    Bouakaz, Saida
    Boufama, Boubakeur
    Guillou, Erwan
    [J]. PATTERN RECOGNITION, 2014, 47 (01) : 238 - 247
  • [4] Chen H., 2013, INT C OPT INSTR TECH
  • [5] Devanne M, 2013, LECT NOTES COMPUT SC, V8158, P456, DOI 10.1007/978-3-642-41190-8_49
  • [6] Depth-images-based pose estimation using regression forests and graphical models
    He, Li
    Wang, Guijin
    Liao, Qingmin
    Xue, Jing-Hao
    [J]. NEUROCOMPUTING, 2015, 164 : 210 - 219
  • [7] Jia XF, 2012, INT C PATT RECOG, P3001
  • [8] Action recognition on motion capture data using a dynemes and forward differences representation
    Kapsouras, Ioannis
    Nikolaidis, Nikos
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2014, 25 (06) : 1432 - 1445
  • [9] Application on Integration Technology of Visualized Hierarchical Information
    Li, Weibo
    He, Yang
    [J]. 2010 THE 3RD INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND INDUSTRIAL APPLICATION (PACIIA2010), VOL I, 2010, : 9 - 12
  • [10] Optimized stereo matching in binocular three-dimensional measurement system using structured light
    Liu, Kun
    Zhou, Changhe
    Wei, Shengbin
    Wang, Shaoqing
    Fan, Xin
    Ma, Jianyong
    [J]. APPLIED OPTICS, 2014, 53 (26) : 6083 - 6090