Research on Human Upper Limb Action Recognition Method Based on Multimodal Heterogeneous Spatial Temporal Graph Network

被引:0
|
作者
Ci, Zelin [1 ]
Ren, Huizhao [1 ]
Liu, Jinming [1 ]
Xie, Songyun [2 ]
Wang, Wendong [1 ,3 ]
机构
[1] Northwestern Polytech Univ, Sch Mech Engn, Xian 710072, Peoples R China
[2] Northwestern Polytech Univ, Elect Informat Coll, Xian 710072, Peoples R China
[3] Sanhang Civil Mil Integrat Innovat Res Inst, Dongguan 523429, Peoples R China
来源
INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2024, PT X | 2025年 / 15210卷
关键词
Action Recognition; Graph Neural Network; Temporal Features; Heterogeneous Graph;
D O I
10.1007/978-981-96-0786-0_23
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph convolutional neural networks have been increasingly used in human action recognition because of their powerful ability to deal with spatial topological relations. Like human action, upper limb action contains spatial and temporal features. For temporal features, graph convolutional networks cannot extract well enough to realize the coupling of spatial-temporal features. This paper proposes a multimodal heterogeneous spatial-temporal network (MHST-GCN) model based on multimodal information. Firstly, the model introduces a temporal graph based on a hybrid sparsity strategy, which captures local and global temporal features in the sequence of human upper limb actions while ensuring computational efficiency. Then, a heterogeneous graph model is proposed for fusing the two modal information to enhance the robustness of the model. Finally, extensive experiments are conducted on two standard datasets, NTU-RGB+D, and a homemade upper limb action dataset. The experimental results demonstrate the effectiveness of the proposed method.
引用
收藏
页码:304 / 318
页数:15
相关论文
共 50 条
  • [1] Spatial Temporal Graph Deconvolutional Network for Skeleton-Based Human Action Recognition
    Peng, Wei
    Shi, Jingang
    Zhao, Guoying
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 244 - 248
  • [2] Skeletal Spatial-Temporal Semantics Guided Homogeneous-Heterogeneous Multimodal Network for Action Recognition
    Zhang, Chenwei
    Hu, Yuxuan
    Yang, Min
    Li, Chengming
    Hu, Xiping
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 3657 - 3666
  • [3] Spatial–Temporal gated graph attention network for skeleton-based action recognition
    Mrugendrasinh Rahevar
    Amit Ganatra
    Pattern Analysis and Applications, 2023, 26 (3) : 929 - 939
  • [4] Spatial Graph Convolutional and Temporal Involution Network for Skeleton-based Action Recognition
    Wan, Huifan
    Pan, Guanghui
    Chen, Yu
    Ding, Danni
    Zou, Maoyang
    PROCEEDINGS OF ACM TURING AWARD CELEBRATION CONFERENCE, ACM TURC 2021, 2021, : 204 - 209
  • [5] Spatial-Temporal Dynamic Graph Attention Network for Skeleton-Based Action Recognition
    Rahevar, Mrugendrasinh
    Ganatra, Amit
    Saba, Tanzila
    Rehman, Amjad
    Bahaj, Saeed Ali
    IEEE ACCESS, 2023, 11 : 21546 - 21553
  • [6] Spatial-Temporal Adaptive Graph Convolutional Network for Skeleton-Based Action Recognition
    Hang, Rui
    Li, MinXian
    COMPUTER VISION - ACCV 2022, PT IV, 2023, 13844 : 172 - 188
  • [7] Spatial-Temporal gated graph attention network for skeleton-based action recognition
    Rahevar, Mrugendrasinh
    Ganatra, Amit
    PATTERN ANALYSIS AND APPLICATIONS, 2023, 26 (03) : 929 - 939
  • [8] An improved spatial temporal graph convolutional network for robust skeleton-based action recognition
    Yuling Xing
    Jia Zhu
    Yu Li
    Jin Huang
    Jinlong Song
    Applied Intelligence, 2023, 53 : 4592 - 4608
  • [9] An improved spatial temporal graph convolutional network for robust skeleton-based action recognition
    Xing, Yuling
    Zhu, Jia
    Li, Yu
    Huang, Jin
    Song, Jinlong
    APPLIED INTELLIGENCE, 2023, 53 (04) : 4592 - 4608
  • [10] STARS: Spatial Temporal Graph Convolution Network for Action Recognition System on FPGAs
    Pei, Songwen
    Wang, Xianrong
    Qin, Wei
    Liang, Sheng
    2021 IEEE 45TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE (COMPSAC 2021), 2021, : 1469 - 1474