FEXNet: Foreground Extraction Network for Human Action Recognition

被引:29
|
作者
Shen, Zhongwei [1 ]
Wu, Xiao-Jun [1 ]
Xu, Tianyang [1 ,2 ]
机构
[1] Jiangnan Univ, Sch Artificial Intelligence & Comp Sci, Wuxi 214122, Jiangsu, Peoples R China
[2] Univ Surrey, Ctr Vis Speech & Signal Proc, Guildford GU2 7XH, Surrey, England
基金
中国国家自然科学基金;
关键词
Convolutional neural networks; Spatiotemporal phenomena; Feature extraction; Three-dimensional displays; Solid modeling; Iron; Image recognition; Foreground-related features; spatiotemporal modeling; action recognition;
D O I
10.1109/TCSVT.2021.3103677
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
As most human actions in video sequences embody the continuous interactions between foregrounds rather than the background scene, it is significant to disentangle these foregrounds from the background for advanced action recognition systems. In this paper, therefore, we propose a Foreground EXtraction (FEX) block to explicitly model the foreground clues to achieve effective management of action subjects. In particular, the designed FEX block contains two components. The first part is a Foreground Enhancement (FE) module, which highlights the potential feature channels related to the action attributes, providing channel-level refinement for the following spatiotemporal modeling. The second phase is a Scene Segregation (SS) module, which splits feature maps into foreground and background. Specifically, a temporal model with dynamic enhancement is constructed for the foreground part, reflecting the essential nature of the action category. While the background is modeled using simple spatial convolutions, mapping the inputs to the consistent feature space. The FEX blocks can be inserted into existing 2D CNNs (denoted as FEXNet) for spatiotemporal modeling, concentrating on the foreground clues for effective action inference. Our experiments performed on Something-Something V1, V2 and Kinetics400 verify the effectiveness of the proposed method.
引用
收藏
页码:3141 / 3151
页数:11
相关论文
共 50 条
  • [21] Human action recognition in complex live videos using graph convolutional network*
    Bharathi, A.
    Sridevi, M.
    COMPUTERS & ELECTRICAL ENGINEERING, 2023, 110
  • [22] Spatial-Temporal Dynamic Graph Attention Network for Skeleton-Based Action Recognition
    Rahevar, Mrugendrasinh
    Ganatra, Amit
    Saba, Tanzila
    Rehman, Amjad
    Bahaj, Saeed Ali
    IEEE ACCESS, 2023, 11 : 21546 - 21553
  • [23] MLENet: Multi-Level Extraction Network for video action recognition
    Wang, Fan
    Li, Xinke
    Xiong, Han
    Mo, Haofan
    Li, Yongming
    PATTERN RECOGNITION, 2024, 154
  • [24] Spatial-Temporal Interleaved Network for Efficient Action Recognition
    Jiang, Shengqin
    Zhang, Haokui
    Qi, Yuankai
    Liu, Qingshan
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2025, 21 (01) : 178 - 187
  • [25] Improved Shift Graph Convolutional Network for Action Recognition With Skeleton
    Li, Chuankun
    Li, Shuai
    Gao, Yanbo
    Guo, Lina
    Li, Wanqing
    IEEE SIGNAL PROCESSING LETTERS, 2023, 30 : 438 - 442
  • [26] View-Adaptive Graph Neural Network for Action Recognition
    Shahid, Ali Raza
    Nawaz, Mehmood
    Fan, Xinqi
    Yan, Hong
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (02) : 969 - 978
  • [27] Human Action Recognition Based on Transfer Learning Approach
    Abdulazeem, Yousry
    Balaha, Hossam Magdy
    Bahgat, Waleed M.
    Badawy, Mahmoud
    IEEE ACCESS, 2021, 9 : 82058 - 82069
  • [28] Pseudo 3D Pose Recognition Network
    Xie, Yuanfeng
    Yu, Xiangyang
    Hong, Weibin
    Xin, Zhaolong
    Chen, Yanwen
    IEEE ACCESS, 2023, 11 : 56380 - 56391
  • [29] Multi-Level Temporal Dilated Dense Prediction for Action Recognition
    Wang, Jinpeng
    Lin, Yiqi
    Zhang, Manlin
    Gao, Yuan
    Ma, Andy J.
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 2553 - 2566
  • [30] STGL-GCN: Spatial-Temporal Mixing of Global and Local Self-Attention Graph Convolutional Networks for Human Action Recognition
    Xie, Zhenggui
    Zheng, Gengzhong
    Miao, Liming
    Huang, Wei
    IEEE ACCESS, 2023, 11 : 16526 - 16532