Multi-level channel attention excitation network for human action recognition in videos

被引:4
|
作者
Wu, Hanbo [1 ]
Ma, Xin [1 ]
Li, Yibin [1 ]
机构
[1] Shandong Univ, Ctr Robot, Sch Control Sci & Engn, Jinan, Peoples R China
基金
中国国家自然科学基金;
关键词
Human action recognition; 2D CNNs; Channel attention; Spatiotemporal modeling;
D O I
10.1016/j.image.2023.116940
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Channel attention mechanism has continuously attracted strong interests and shown great potential in enhancing the performance of deep CNNs. However, when applied to video-based human action recognition task, most existing methods generally learn channel attention at frame level, which ignores the temporal dependencies and may limit the recognition performance. In this paper, we propose a novel multi-level channel attention excitation (MCAE) module to model the temporal-related channel attention at both frame and video levels. Specifically, based on video convolutional feature maps, frame-level channel attention (FCA) is generated by exploring time-channel correlations, and video-level channel attention (VCA) is generated by aggregating global motion variations. MCAE firstly recalibrates video feature responses with frame-wise FCA, and then activates the motion-sensitive channel features with motion-aware VCA. MCAE module learns the channel discriminability from multiple levels and can act as a guidance to facilitate efficient spatiotemporal feature modeling in activated motion-sensitive channels. It can be flexibly embedded into 2D networks with very limited extra computation cost to construct MCAE-Net, which effectively enhances the spatiotemporal representation of 2D models for video action recognition task Extensive experiments on five human action datasets show that our method achieves superior or very competitive performance compared with the state -of-the-arts, which demonstrates the effectiveness of the proposed method for improving the performance of human action recognition.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Multi-level feature fusion capsule network with self-attention for facial expression recognition
    Huang, Zhiji
    Yu, Songsen
    Liang, Jun
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (02)
  • [32] Multi-Level Context Aggregation Network with Channel-Wise Attention for Salient Object Detection
    Jia, Zihui
    Weng, Zhenyu
    Wan, Fang
    Zhu, Yuesheng
    IEEE Access, 2020, 8 : 102303 - 102312
  • [33] Multi-Level Context Aggregation Network With Channel-Wise Attention for Salient Object Detection
    Jia, Zihui
    Weng, Zhenyu
    Wan, Fang
    Zhu, Yuesheng
    IEEE ACCESS, 2020, 8 : 102303 - 102312
  • [34] Human Action Recognition in Unconstrained Trimmed Videos Using Residual Attention Network and Joints Path Signature
    Ahmad, Tasweer
    Jin, Lianwen
    Feng, Jialuo
    Tang, Guozhi
    IEEE ACCESS, 2019, 7 : 121212 - 121222
  • [35] Multi-STMT: Multi-Level Network for Human Activity Recognition Based on Wearable Sensors
    Zhang, Haoran
    Xu, Linhai
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 12
  • [36] Multi-level Motion Attention for Human Motion Prediction
    Mao, Wei
    Liu, Miaomiao
    Salzmann, Mathieu
    Li, Hongdong
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (09) : 2513 - 2535
  • [37] Human Action Recognition with Multi-Level Granularity and Pair-wise Hyper GCN
    Alsarhan, Tamam
    Ali, Syed Sadaf
    Alsarhan, Ayoub
    Ganapathi, Iyyakutti Iyappan
    Werghi, Naoufel
    2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024, 2024,
  • [38] Multi-level Motion Attention for Human Motion Prediction
    Wei Mao
    Miaomiao Liu
    Mathieu Salzmann
    Hongdong Li
    International Journal of Computer Vision, 2021, 129 : 2513 - 2535
  • [39] Multi-Temporal Convolutions for Human Action Recognition in Videos
    Stergiou, Alexandros
    Poppe, Ronald
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [40] Multi-level spatial attention network for image data segmentation
    Guo, Jun
    Jiang, Zhixiong
    Jiang, Dingchao
    INTERNATIONAL JOURNAL OF EMBEDDED SYSTEMS, 2021, 14 (03) : 289 - 299