Skeleton-based human action recognition by fusing attention based three-stream convolutional neural network and SVM

被引:0
|
作者
Ren, Fang [1 ]
Tang, Chao [1 ]
Tong, Anyang [1 ]
Wang, Wenjian [2 ]
机构
[1] Hefei Univ, Sch Artificial Intelligence & Big Data, Hefei, Peoples R China
[2] Shanxi Univ, Sch Comp & Informat Technol, Taiyuan, Peoples R China
关键词
Skeleton-based human action recognition; Convolutional neural network; Attention mechanism; Support vector machine; Spatial-temporal feature; RECOMMENDATION SYSTEM; VISION;
D O I
10.1007/s11042-023-15334-9
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This work proposes a method, aiming the 3D skeleton sequence, for the human action recognition by fusing the attention-based three-stream convolutional neural network and support vector machine. The traditional action recognition methods primarily employ RGB video as input. However, RGB video has issues with respect to large data volume, low semanticity, and ease of making the model interfered by irrelevant information such as the background. The efficient and advanced human action information contained in the 3D skeleton sequence facilitates human behavior recognition. First, the information of 3D coordinates, temporal-difference information, and spatial-difference information of joints are extracted from the raw skeleton data, and the above information is input into the respective convolutional neural networks for pre-training. Then, the pre-trained network model extracts the feature containing the spatial-temporal information. Finally, the mixed feature vectors are input into the support vector machine for training and classification. Under the X-View and X-Sub benchmarks, the accuracy on the open dataset NTU RGB+D is 92.6% and 86.7% respectively, demonstrating that the method proposed for incorporating multistream feature learning, feature fusing, and hybrid model can improve the recognition accuracy.
引用
收藏
页码:6273 / 6295
页数:23
相关论文
共 50 条
  • [1] Skeleton-based human action recognition by fusing attention based three-stream convolutional neural network and SVM
    Fang Ren
    Chao Tang
    Anyang Tong
    Wenjian Wang
    Multimedia Tools and Applications, 2024, 83 : 6273 - 6295
  • [2] Two Stream Multi-Attention Graph Convolutional Network for Skeleton-Based Action Recognition
    Zhou, Huijian
    Tian, Zhiqiang
    Du, Shaoyi
    ARTIFICIAL INTELLIGENCE AND ROBOTICS, ISAIR 2023, 2024, 1998 : 112 - 120
  • [3] Skeleton-Based Square Grid for Human Action Recognition With 3D Convolutional Neural Network
    Ding, Wenwen
    Ding, Chongyang
    Li, Guang
    Liu, Kai
    IEEE ACCESS, 2021, 9 : 54078 - 54089
  • [4] Sequence Segmentation Attention Network for Skeleton-Based Action Recognition
    Zhang, Yujie
    Cai, Haibin
    ELECTRONICS, 2023, 12 (07)
  • [5] Prompt-supervised dynamic attention graph convolutional network for skeleton-based action recognition
    Zhu, Shasha
    Sun, Lu
    Ma, Zeyuan
    Li, Chenxi
    He, Dongzhi
    NEUROCOMPUTING, 2025, 611
  • [6] Attention adjacency matrix based graph convolutional networks for skeleton-based action recognition
    Xie, Jun
    Miao, Qiguang
    Liu, Ruyi
    Xin, Wentian
    Tang, Lei
    Zhong, Sheng
    Gao, Xuesong
    NEUROCOMPUTING, 2021, 440 (440) : 230 - 239
  • [7] Insight on Attention Modules for Skeleton-Based Action Recognition
    Jiang, Quanyan
    Wu, Xiaojun
    Kittler, Josef
    PATTERN RECOGNITION AND COMPUTER VISION, PT I, 2021, 13019 : 242 - 255
  • [8] Temporal-Channel Attention and Convolution Fusion for Skeleton-Based Human Action Recognition
    Liang, Chengwu
    Yang, Jie
    Du, Ruolin
    Hu, Wei
    Hou, Ning
    IEEE ACCESS, 2024, 12 : 64937 - 64948
  • [9] Skeleton Action Recognition Based on Multi-Stream Spatial Attention Graph Convolutional SRU Network
    Zhao J.-N.
    She Q.-S.
    Meng M.
    Chen Y.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2022, 50 (07): : 1579 - 1585
  • [10] Spatiotemporal Graph Autoencoder Network for Skeleton-Based Human Action Recognition
    Abduljalil, Hosam
    Elhayek, Ahmed
    Marish Ali, Abdullah
    Alsolami, Fawaz
    AI, 2024, 5 (03) : 1695 - 1708