Skeleton-Based Multifeatures and Multistream Network for Real-Time Action Recognition

被引:13
作者
Deng, Zhiwen [1 ,2 ]
Gao, Qing [1 ]
Ju, Zhaojie [3 ]
Yu, Xiang [4 ]
机构
[1] Sun Yat Sen Univ, Sch Elect & Commun Engn, Shenzhen 518107, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Sch Telecommun & Informat Engn, Chongqing 400065, Peoples R China
[3] Univ Portsmouth, Sch Comp, Portsmouth PO1 3HE, Hants, England
[4] Chongqing Univ Posts & Telecommun, Sch Commun & Informat Engn, Chongqing 400065, Peoples R China
基金
中国国家自然科学基金;
关键词
Skeleton; Feature extraction; Sensors; Real-time systems; Human-robot interaction; Cameras; Face recognition; Human-computer interaction; multifeature; real-time; skeleton-based action recognition;
D O I
10.1109/JSEN.2023.3246133
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Action recognition is a hot topic in the field of computer vision. It has been widely used in human-computer/robot interaction, abnormal behavior monitoring, and medical assistive. Because of the excellent robustness of skeleton data, it has attracted many scholars to research skeleton-based action recognition. Most of the current skeleton-based action recognition methods suffer from the incomplete and poor generalization of the input features, inadequate feature extraction by the network model, and an imbalance between recognition accuracy and model size. We analyze the critical skeleton features for action recognition to solve these problems and propose a multifeatures and multistream network (MM-Net) for real-time action recognition. First, three pairs of features are proposed, which are the joint distance (JD) and JD velocity (JDV), joint angle (JA) and JA velocity (JAV), and fast-action joint position (FJP) and slow-action joint position (SJP). Second, an MM-Net is proposed by using a 1-D convolutional neural network (1DCNN) to reduce the number of parameters of the model and fully extract the three pairs of features. As a result, MM-Net achieves the highest accuracies on both JHMDB (86.5%) and SHREC (96.4% on coarse and 93.3% on fine datasets). In addition, MM-Net is applied to a human-robot interaction (HRI) platform, which proves the practicality of MM-Net.
引用
收藏
页码:7397 / 7409
页数:13
相关论文
共 41 条
  • [1] [Anonymous], 2017, P STAG
  • [2] Learning a 3D Human Pose Distance Metric from Geometric Pose Descriptor
    Chen, Cheng
    Zhuang, Yueting
    Nie, Feiping
    Yang, Yi
    Wu, Fei
    Xiao, Jun
    [J]. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2011, 17 (11) : 1676 - 1689
  • [3] MFA-Net: Motion Feature Augmented Network for Dynamic Hand Gesture Recognition from Skeletal Data
    Chen, Xinghao
    Wang, Guijin
    Guo, Hengkai
    Zhang, Cairong
    Wang, Hang
    Zhang, Li
    [J]. SENSORS, 2019, 19 (02)
  • [4] PoTion: Pose MoTion Representation for Action Recognition
    Choutas, Vasileios
    Weinzaepfel, Philippe
    Revaud, Jerome
    Schmid, Cordelia
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7024 - 7033
  • [5] De Smedt Q., 2017, 3DOR 10 EUR WORKSH 3, P1, DOI DOI 10.2312/3DOR.20171049
  • [6] Skeleton-based Dynamic hand gesture recognition
    De Smedt, Quentin
    Wannous, Hazem
    Vandeborre, Jean-Philippe
    [J]. PROCEEDINGS OF 29TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, (CVPRW 2016), 2016, : 1206 - 1214
  • [7] Devineau G., 2018, RECONNAISSANCE FORME
  • [8] REConvertor: Transforming Textual Use Cases to High-Level Message Sequence Chart
    Ding, Zuohua
    Shuai, Tiantian
    Jiang, Mingyue
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY AND SECURITY COMPANION (QRS-C), 2017, : 610 - 611
  • [9] SlowFast Networks for Video Recognition
    Feichtenhofer, Christoph
    Fan, Haoqi
    Malik, Jitendra
    He, Kaiming
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6201 - 6210
  • [10] An Efficient RGB-D Hand Gesture Detection Framework for Dexterous Robot Hand-Arm Teleoperation System
    Gao, Qing
    Ju, Zhaojie
    Chen, Yongquan
    Wang, Qiwen
    Chi, Chuliang
    [J]. IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2023, 53 (01) : 13 - 23