Research on Human Upper Limb Action Recognition Method Based on Multimodal Heterogeneous Spatial Temporal Graph Network

被引:0
|
作者
Ci, Zelin [1 ]
Ren, Huizhao [1 ]
Liu, Jinming [1 ]
Xie, Songyun [2 ]
Wang, Wendong [1 ,3 ]
机构
[1] Northwestern Polytech Univ, Sch Mech Engn, Xian 710072, Peoples R China
[2] Northwestern Polytech Univ, Elect Informat Coll, Xian 710072, Peoples R China
[3] Sanhang Civil Mil Integrat Innovat Res Inst, Dongguan 523429, Peoples R China
来源
INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2024, PT X | 2025年 / 15210卷
关键词
Action Recognition; Graph Neural Network; Temporal Features; Heterogeneous Graph;
D O I
10.1007/978-981-96-0786-0_23
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph convolutional neural networks have been increasingly used in human action recognition because of their powerful ability to deal with spatial topological relations. Like human action, upper limb action contains spatial and temporal features. For temporal features, graph convolutional networks cannot extract well enough to realize the coupling of spatial-temporal features. This paper proposes a multimodal heterogeneous spatial-temporal network (MHST-GCN) model based on multimodal information. Firstly, the model introduces a temporal graph based on a hybrid sparsity strategy, which captures local and global temporal features in the sequence of human upper limb actions while ensuring computational efficiency. Then, a heterogeneous graph model is proposed for fusing the two modal information to enhance the robustness of the model. Finally, extensive experiments are conducted on two standard datasets, NTU-RGB+D, and a homemade upper limb action dataset. The experimental results demonstrate the effectiveness of the proposed method.
引用
收藏
页码:304 / 318
页数:15
相关论文
共 50 条
  • [21] A SPATIAL-TEMPORAL CONSTRAINT-BASED ACTION RECOGNITION METHOD
    Han, Tingting
    Yao, Hongxun
    Zhang, Yanhao
    Xu, Pengfei
    2013 20TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2013), 2013, : 2767 - 2771
  • [22] Spatial-temporal graph attention networks for skeleton-based action recognition
    Huang, Qingqing
    Zhou, Fengyu
    He, Jiakai
    Zhao, Yang
    Qin, Runze
    JOURNAL OF ELECTRONIC IMAGING, 2020, 29 (05)
  • [23] STGNN-LMR: A Spatial-Temporal Graph Neural Network Approach Based on sEMG Lower Limb Motion Recognition
    Mao, Weifan
    Ma, Bin
    Li, Zhao
    Zhang, Jianxing
    Lu, Yizhou
    Yu, Zhuting
    Zhang, Feng
    JOURNAL OF BIONIC ENGINEERING, 2023, 21 (1): : 256 - 269
  • [24] Spatial-temporal pyramid based Convolutional Neural Network for action recognition
    Zheng, Zhenxing
    An, Gaoyun
    Wu, Dapeng
    Ruan, Qiuqi
    NEUROCOMPUTING, 2019, 358 : 446 - 455
  • [25] Graph transformer network with temporal kernel attention for skeleton-based action recognition
    Liu, Yanan
    Zhang, Hao
    Xu, Dan
    He, Kangjian
    KNOWLEDGE-BASED SYSTEMS, 2022, 240
  • [26] Temporal Receptive Field Graph Convolutional Network for Skeleton-based Action Recognition
    Zhang, Qingqi
    Wu, Ren
    Nakata, Mitsuru
    Ge, Qi-Wei
    2024 INTERNATIONAL TECHNICAL CONFERENCE ON CIRCUITS/SYSTEMS, COMPUTERS, AND COMMUNICATIONS, ITC-CSCC 2024, 2024,
  • [27] An attentional spatial temporal graph convolutional network with co-occurrence feature learning for action recognition
    Dong Tian
    Zhe-Ming Lu
    Xiao Chen
    Long-Hua Ma
    Multimedia Tools and Applications, 2020, 79 : 12679 - 12697
  • [28] Multi-granular spatial-temporal synchronous graph convolutional network for robust action recognition
    Li, Chang
    Huang, Qian
    Mao, Yingchi
    Li, Xing
    Wu, Jie
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 257
  • [29] Refined Spatial Network for Human Action Recognition
    Wu, Chunlei
    Cao, Haiwen
    Zhang, Weishan
    Wang, Leiquan
    Wei, Yiwei
    Peng, Zexin
    IEEE ACCESS, 2019, 7 : 111043 - 111052
  • [30] An attentional spatial temporal graph convolutional network with co-occurrence feature learning for action recognition
    Tian, Dong
    Lu, Zhe-Ming
    Chen, Xiao
    Ma, Long-Hua
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (17-18) : 12679 - 12697