Depth-Based vs. Color-Based Pose Estimation in Human Action Recognition

被引:0
作者
Malawski, Filip [1 ]
Jankowski, Bartosz [1 ]
机构
[1] AGH Univ Sci & Technol, Inst Comp Sci, Krakow, Poland
来源
ADVANCES IN VISUAL COMPUTING, ISVC 2022, PT I | 2022年 / 13598卷
关键词
Pose estimation; Human action recognition; Depth modality; Color modality; Kinect;
D O I
10.1007/978-3-031-20713-6_26
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Recent advances in deep learning resulted in the emergence of accurate models for human pose estimation in color videos. Distance between automatically estimated and manually annotated joint positions is commonly used for the evaluation of such methods. However, from a practical point of view, pose estimation is not a goal by itself. Therefore, in this work, we study how useful are state-of-the-art deep learning pose estimation approaches in a practical scenario of human action recognition. We compare different variants of pose estimation models with the baseline provided by the Kinect skeleton tracking, which, until recently, was the most widely used solution in such applications. We present a comprehensive framework for pose-based action recognition evaluation, which consists of both classical machine learning approaches, including feature extraction, selection, and classification steps, as well as more recent end-to-end methods. Extensive evaluation on four publicly available datasets shows, that by using state-of-the-art neural network models for pose tracking, color-based action recognition matches, or even outperforms, that of the depth-based one.
引用
收藏
页码:336 / 346
页数:11
相关论文
共 23 条
[1]   TSFEL: Time Series Feature Extraction Library [J].
Barandas, Marilia ;
Folgado, Duarte ;
Fernandes, Leticia ;
Santos, Sara ;
Abreu, Mariana ;
Bota, Patricia ;
Liu, Hui ;
Schultz, Tanja ;
Gamboa, Hugo .
SOFTWAREX, 2020, 11
[2]  
Bazarevsky V, 2020, Arxiv, DOI [arXiv:2006.10204, DOI 10.48550/ARXIV.2006.10204]
[3]   Action Recognition Using Rate-Invariant Analysis of Skeletal Shape Trajectories [J].
Ben Amor, Boulbaba ;
Su, Jingyong ;
Srivastava, Anuj .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (01) :1-13
[4]   A survey of depth and inertial sensor fusion for human action recognition [J].
Chen, Chen ;
Jafari, Roozbeh ;
Kehtarnavaz, Nasser .
MULTIMEDIA TOOLS AND APPLICATIONS, 2017, 76 (03) :4405-4425
[5]  
Chen C, 2015, IEEE IMAGE PROC, P168, DOI 10.1109/ICIP.2015.7350781
[6]   Human Activity Recognition Process Using 3-D Posture Data [J].
Gaglio, Salvatore ;
Lo Re, Giuseppe ;
Morana, Marco .
IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2015, 45 (05) :586-597
[7]   Extremely randomized trees [J].
Geurts, P ;
Ernst, D ;
Wehenkel, L .
MACHINE LEARNING, 2006, 63 (01) :3-42
[8]   DensePose: Dense Human Pose Estimation In The Wild [J].
Guler, Riza Alp ;
Neverova, Natalia ;
Kokkinos, Lasonas .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7297-7306
[9]   An overview of Human Action Recognition in sports based on Computer Vision [J].
Host, Kristina ;
Ivasic-Kos, Marina .
HELIYON, 2022, 8 (06)
[10]   A New Representation of Skeleton Sequences for 3D Action Recognition [J].
Ke, Qiuhong ;
Bennamoun, Mohammed ;
An, Senjian ;
Sohel, Ferdous ;
Boussaid, Farid .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :4570-4579