CNN-based and DTW features for human activity recognition on depth maps

被引:13
作者
Trelinski, Jacek [1 ]
Kwolek, Bogdan [1 ]
机构
[1] AGH Univ Sci & Technol, Dept Comp Sci, 30 Mickiewicza Av,Bldg D-17, PL-30059 Krakow, Poland
关键词
Convolutional neural networks; Multivariate time-series; Ensembles; Depth-based human action recognition;
D O I
10.1007/s00521-021-06097-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we present a new algorithm for human action recognition on raw depth maps. At the beginning, for each class we train a separate one-against-all convolutional neural network (CNN) to extract class-specific features representing person shape. Each class-specific, multivariate time-series is processed by a Siamese multichannel 1D CNN or a multichannel 1D CNN to determine features representing actions. Afterwards, for the nonzero pixels representing the person shape in each depth map we calculate statistical features. On multivariate time-series of such features we determine Dynamic Time Warping (DTW) features. They are determined on the basis of DTW distances between all training time-series. Finally, each class-specific feature vector is concatenated with the DTW feature vector. For each action category we train a multiclass classifier, which predicts probability distribution of class labels. From pool of such classifiers we select a number of classifiers such that an ensemble built on them achieves the best classification accuracy. Action recognition is performed by a soft voting ensemble that averages distributions calculated by such classifiers with the largest discriminative power. We demonstrate experimentally that on MSR-Action3D and UTD-MHAD datasets the proposed algorithm attains promising results and outperforms several state-of-the-art depth-based algorithms.
引用
收藏
页码:14551 / 14563
页数:13
相关论文
共 24 条
  • [1] Chen C, 2015, IEEE IMAGE PROC, P168, DOI 10.1109/ICIP.2015.7350781
  • [2] Learning a similarity metric discriminatively, with application to face verification
    Chopra, S
    Hadsell, R
    LeCun, Y
    [J]. 2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, : 539 - 546
  • [3] SOFMLS: Online Self-Organizing Fuzzy Modified Least-Squares Network
    de Jesus Rubio, Jose
    [J]. IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2009, 17 (06) : 1296 - 1309
  • [4] Hadsell R., 2006, IEEE C COMP VIS PATT, V2, P1735
  • [5] Skeleton Optical Spectra-Based Action Recognition Using Convolutional Neural Networks
    Hou, Yonghong
    Li, Zhaoyang
    Wang, Pichao
    Li, Wanqing
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2018, 28 (03) : 807 - 811
  • [6] Deep Convolutional Neural Networks for Human Action Recognition Using Depth Maps and Postures
    Kamel, Aouaidjia
    Sheng, Bin
    Yang, Po
    Li, Ping
    Shen, Ruimin
    Feng, David Dagan
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2019, 49 (09): : 1806 - 1819
  • [7] Koch G., 2015, P ICML DEEP LEARN WO, VVolume 2
  • [8] Efficient Visual Classification by Fuzzy Rules
    Korytkowski, Marcin
    Scherer, Rafal
    Szajerman, Dominik
    Polap, Dawid
    Wozniak, Marcin
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ-IEEE), 2020,
  • [9] Application on Integration Technology of Visualized Hierarchical Information
    Li, Weibo
    He, Yang
    [J]. 2010 THE 3RD INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND INDUSTRIAL APPLICATION (PACIIA2010), VOL I, 2010, : 9 - 12
  • [10] Liang Bo, 2015, 2015 IEEE Transportation Electrification Conference and Expo (ITEC), P1, DOI 10.1109/ITEC.2015.7165751