Effective human action recognition using global and local offsets of skeleton joints

被引:13
作者
Sun, Bin [1 ]
Kong, Dehui [1 ]
Wang, Shaofan [1 ]
Wang, Lichun [1 ]
Wang, Yuping [1 ]
Yin, Baocai [2 ]
机构
[1] Beijing Univ Technol, BJUT Fac Informat Technol, Beijing Adv Innovat Ctr Future Internet Technol, Beijing Key Lab Multimedia & Intelligent Software, Beijing 100124, Peoples R China
[2] Dalian Univ Technol, Coll Comp Sci & Technol, Fac Elect Informat & Elect Engn, Dalian 116024, Peoples R China
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
Action recognition; Skeleton joints; Offsets; Histogram representation; Naive-Bayes-Nearest-Neighbor; ENSEMBLE; FEATURES; SPACE;
D O I
10.1007/s11042-018-6370-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Human action recognition based on 3D skeleton joints is an important yet challenging task. While many research work are devoted to 3D action recognition, they mainly suffer from two problems: complex model representation and low implementation efficiency. To tackle these problems, we propose an effective and efficient framework for 3D action recognition using a global-and-local histogram representation model. Our method consists of a global-and-local featuring phase, a saturation based histogram representation phase, and a classification phase. The global-and-local featuring phase captures the global feature and local feature of each action sequence using the joint displacement between the current frame and the first frame, and the joint displacement between pairwise fixed-skip frames, respectively. The saturation based histogram representation phase captures the histogram representation of each joint considering the motion independence of joints and saturation of each histogram bin. The classification phase measures the distance of each joint histogram-to-class. Besides, we produce a novel action dataset called BJUT Kinect dataset, which consists of multi-period motion clips and intra-class variations. We compare our method with many state-of-the-art methods on BJUT Kinect dataset, UCF Kinect dataset, Florence 3D action dataset, MSR-Action3D dataset, and NTU RGB+D Dataset. The results show that our method achieves both higher accuracy and efficiency for 3D action recognition.
引用
收藏
页码:6329 / 6353
页数:25
相关论文
共 58 条
  • [1] Improving bag-of-poses with semi-temporal pose descriptors for skeleton-based action recognition
    Agahian, Saeid
    Negin, Farhood
    Kose, Cemal
    [J]. VISUAL COMPUTER, 2019, 35 (04) : 591 - 607
  • [2] Evolutionary joint selection to improve human action recognition with RGB-D devices
    Andre Chaaraoui, Alexandros
    Ramon Padilla-Lopez, Jose
    Climent-Perez, Pau
    Florez-Revuelta, Francisco
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2014, 41 (03) : 786 - 794
  • [3] Elastic Functional Coding of Riemannian Trajectories
    Anirudh, Rushil
    Turaga, Pavan
    Su, Jingyong
    Srivastava, Anuj
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (05) : 922 - 936
  • [4] [Anonymous], 2008, P 2008 IEEE C COMPUT, DOI DOI 10.1109/CVPR.2008.4587735
  • [5] [Anonymous], 2015, PROC CVPR IEEE
  • [6] Hidden Markov Model on a unit hypersphere space for gesture trajectory recognition
    Beh, Jounghoon
    Han, David K.
    Durasiwami, Ramani
    Ko, Hanseok
    [J]. PATTERN RECOGNITION LETTERS, 2014, 36 : 144 - 153
  • [7] In defense of Nearest-Neighbor based image classification
    Boiman, Oren
    Shechtman, Eli
    Irani, Michal
    [J]. 2008 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-12, 2008, : 1992 - +
  • [8] A novel hierarchical framework for human action recognition
    Chen, Hongzhao
    Wang, Guijin
    Xue, Jing-Hao
    He, Li
    [J]. PATTERN RECOGNITION, 2016, 55 : 148 - 159
  • [9] TriViews: A general framework to use 3D depth data effectively for action recognition
    Chen, Wenbin
    Guo, Guodong
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2015, 26 : 182 - 191
  • [10] Integrated video object tracking with applications in trajectory-based event detection
    Cheng, Hsu-Yung
    Hwang, Jenq-Neng
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2011, 22 (07) : 673 - 685