Skeleton-Based Multifeatures and Multistream Network for Real-Time Action Recognition

被引:13
作者
Deng, Zhiwen [1 ,2 ]
Gao, Qing [1 ]
Ju, Zhaojie [3 ]
Yu, Xiang [4 ]
机构
[1] Sun Yat Sen Univ, Sch Elect & Commun Engn, Shenzhen 518107, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Sch Telecommun & Informat Engn, Chongqing 400065, Peoples R China
[3] Univ Portsmouth, Sch Comp, Portsmouth PO1 3HE, Hants, England
[4] Chongqing Univ Posts & Telecommun, Sch Commun & Informat Engn, Chongqing 400065, Peoples R China
基金
中国国家自然科学基金;
关键词
Skeleton; Feature extraction; Sensors; Real-time systems; Human-robot interaction; Cameras; Face recognition; Human-computer interaction; multifeature; real-time; skeleton-based action recognition;
D O I
10.1109/JSEN.2023.3246133
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Action recognition is a hot topic in the field of computer vision. It has been widely used in human-computer/robot interaction, abnormal behavior monitoring, and medical assistive. Because of the excellent robustness of skeleton data, it has attracted many scholars to research skeleton-based action recognition. Most of the current skeleton-based action recognition methods suffer from the incomplete and poor generalization of the input features, inadequate feature extraction by the network model, and an imbalance between recognition accuracy and model size. We analyze the critical skeleton features for action recognition to solve these problems and propose a multifeatures and multistream network (MM-Net) for real-time action recognition. First, three pairs of features are proposed, which are the joint distance (JD) and JD velocity (JDV), joint angle (JA) and JA velocity (JAV), and fast-action joint position (FJP) and slow-action joint position (SJP). Second, an MM-Net is proposed by using a 1-D convolutional neural network (1DCNN) to reduce the number of parameters of the model and fully extract the three pairs of features. As a result, MM-Net achieves the highest accuracies on both JHMDB (86.5%) and SHREC (96.4% on coarse and 93.3% on fine datasets). In addition, MM-Net is applied to a human-robot interaction (HRI) platform, which proves the practicality of MM-Net.
引用
收藏
页码:7397 / 7409
页数:13
相关论文
共 41 条
  • [21] A Mobile Robot Hand-Arm Teleoperation System by Vision and IMU
    Li, Shuang
    Jiang, Jiaxi
    Ruppel, Philipp
    Liang, Hongzhuo
    Ma, Xiaojian
    Hendrich, Norman
    Sun, Fuchun
    Zhang, Jianwei
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10900 - 10906
  • [22] Adaptive RNN Tree for Large-Scale Human Action Recognition
    Li, Wenbo
    Wen, Longyin
    Chang, Ming-Ching
    Lim, Ser Nam
    Lyu, Siwei
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 1453 - 1461
  • [23] Spatio-Temporal LSTM with Trust Gates for 3D Human Action Recognition
    Liu, Jun
    Shahroudy, Amir
    Xu, Dong
    Wang, Gang
    [J]. COMPUTER VISION - ECCV 2016, PT III, 2016, 9907 : 816 - 833
  • [24] Enhanced skeleton visualization for view invariant human action recognition
    Liu, Mengyuan
    Liu, Hong
    Chen, Chen
    [J]. PATTERN RECOGNITION, 2017, 68 : 346 - 362
  • [25] NAGNN: Classification of COVID-19 based on neighboring aware representation from deep graph neural network
    Lu, Siyuan
    Zhu, Ziquan
    Manuel Gorriz, Juan
    Wang, Shui-Hua
    Zhang, Yu-Dong
    [J]. INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (02) : 1572 - 1598
  • [26] Detection of abnormal brain in MRI via improved AlexNet and ELM optimized by chaotic bat algorithm
    Lu, Siyuan
    Wang, Shui-Hua
    Zhang, Yu-Dong
    [J]. NEURAL COMPUTING & APPLICATIONS, 2021, 33 (17) : 10799 - 10811
  • [27] Ludl D, 2019, IEEE INT C INTELL TR, P581, DOI 10.1109/ITSC.2019.8917128
  • [28] Domain and View-Point Agnostic Hand Action Recognition
    Sabater, Alberto
    Alonso, Inigo
    Montesano, Luis
    Murillo, Ana Cristina
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (04) : 7823 - 7830
  • [29] Selvaraju RR, 2020, INT J COMPUT VISION, V128, P336, DOI [10.1007/s11263-019-01228-7, 10.1109/ICCV.2017.74]
  • [30] Pose and Joint-Aware Action Recognition
    Shah, Anshul
    Mishra, Shlok
    Bansal, Ankan
    Chen, Jun-Cheng
    Chellappa, Rama
    Shrivastava, Abhinav
    [J]. 2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 141 - 151