ASMGCN: Attention-Based Semantic-Guided Multistream Graph Convolution Network for Skeleton Action Recognition

被引:1
|
作者
Zhang, Moyan [1 ]
Quan, Zhenzhen [1 ]
Wang, Wei [1 ]
Chen, Zhe [1 ]
Guo, Xiaoshan [1 ]
Li, Yujun [1 ]
机构
[1] Shandong Univ, Sch Informat Sci & Engn, Qingdao 266237, Shandong, Peoples R China
关键词
Skeleton; Feature extraction; Convolution; Sensors; Data models; Data mining; Joints; Attention mechanism; graph convolution network (GCN); multistream network; skeleton-based action recognition;
D O I
10.1109/JSEN.2024.3388154
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In recent years, the field of action recognition using spatio-temporal graph convolution models for human skeletal data has made significant progress. However, current methodologies tend to prioritize spatial graph convolution, which leads to an underutilization of valuable information present in skeletal data. It limits the model's ability to effectively capture complex data patterns, especially in time-series data, ultimately impacting recognition accuracy significantly. To address the above issues, this article introduces an attention-based semantic-guided multistream graph convolution network (ASMGCN), which can extract the deep features in skeletal data more fully. Specifically, ASMGCN incorporates a novel temporal convolutional module featuring an attention mechanism and a multiscale residual network, which can dynamically adjust the weights between skeleton graphs at different time points, enabling better capture of relational features. In addition, semantic information is introduced into the loss function, enhancing the model's ability to distinguish similar actions. Furthermore, the coordinate information of different joints within the same frame is explored to generate new relative position features known as centripetal and centrifugal streams based on the center of gravity. These features are integrated with the original position and motion features of skeleton, including joints and bones, enriching the inputs to the GCN. Experimental results on the NW-UCLA, NTU RGB + D (NTU60), and NTU RGB + D 120 (NTU120) datasets demonstrate that ASMGCN outperforms other state-of-the-art (SOTA) human action recognition (HAR) methods, signifying its potential in advancing the field of action recognition using skeletal data.
引用
收藏
页码:20064 / 20075
页数:12
相关论文
共 50 条
  • [1] Improved semantic-guided network for skeleton-based action recognition
    Mansouri, Amine
    Bakir, Toufik
    Elzaar, Abdellah
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 104
  • [2] Multi-Scale Mixed Dense Graph Convolution Network for Skeleton-Based Action Recognition
    Xia, Hailun
    Gao, Xinkai
    IEEE ACCESS, 2021, 9 (09): : 36475 - 36484
  • [3] SelfGCN: Graph Convolution Network With Self-Attention for Skeleton-Based Action Recognition
    Wu, Zhize
    Sun, Pengpeng
    Chen, Xin
    Tang, Keke
    Xu, Tong
    Zou, Le
    Wang, Xiaofeng
    Tan, Ming
    Cheng, Fan
    Weise, Thomas
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 4391 - 4403
  • [4] AeS-GCN: Attention-enhanced semantic-guided graph convolutional networks for skeleton-based action recognition
    Xu, Qing
    LiU, Feng
    Fu, Ziwang
    Zhou, Aimin
    Qi, Jiayin
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2022, 33 (3-4)
  • [5] Dynamic Semantic-Based Spatial-Temporal Graph Convolution Network for Skeleton-Based Human Action Recognition
    Xie, Jianyang
    Meng, Yanda
    Zhao, Yitian
    Nguyen, Anh
    Yang, Xiaoyun
    Zheng, Yalin
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 6691 - 6704
  • [6] Feedback Graph Convolutional Network for Skeleton-Based Action Recognition
    Yang, Hao
    Yan, Dan
    Zhang, Li
    Sun, Yunda
    Li, Dong
    Maybank, Stephen J.
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 164 - 175
  • [7] Skeleton-Based Multifeatures and Multistream Network for Real-Time Action Recognition
    Deng, Zhiwen
    Gao, Qing
    Ju, Zhaojie
    Yu, Xiang
    IEEE SENSORS JOURNAL, 2023, 23 (07) : 7397 - 7409
  • [8] Temporal-Channel Attention and Convolution Fusion for Skeleton-Based Human Action Recognition
    Liang, Chengwu
    Yang, Jie
    Du, Ruolin
    Hu, Wei
    Hou, Ning
    IEEE ACCESS, 2024, 12 : 64937 - 64948
  • [9] Effective skeleton topology and semantics-guided adaptive graph convolution network for action recognition
    Zhong-Xiang Qiu
    Hong-Bo Zhang
    Wei-Mo Deng
    Ji-Xiang Du
    Qing Lei
    Guo-Liang Zhang
    The Visual Computer, 2023, 39 : 2191 - 2203
  • [10] Effective skeleton topology and semantics-guided adaptive graph convolution network for action recognition
    Qiu, Zhong-Xiang
    Zhang, Hong-Bo
    Deng, Wei-Mo
    Du, Ji-Xiang
    Lei, Qing
    Zhang, Guo-Liang
    VISUAL COMPUTER, 2023, 39 (05) : 2191 - 2203