Recognizing human activities with the use of Convolutional Block Attention Module

被引:1
作者
Zakariah, Mohammed [1 ]
Alnuaim, Abeer [1 ]
机构
[1] King Saud Univ, Coll Appl Studies & Community Serv, Dept Comp Sci & Engn, POB 22459, Riyadh 11495, Saudi Arabia
关键词
Human activity recognition; Human behaviour recognition; Deep-learning; Convolutional Block Attention Module (CBAM); Convolution Neural Network; Spatial Attention Module; HUMAN ACTION RECOGNITION;
D O I
10.1016/j.eij.2024.100536
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Human Activity Recognition (HAR) is crucial for the advancement of applications in smart environments, communication, IoT, security, and healthcare monitoring. Convolutional neural networks (CNNs) have made substantial contributions to human activity recognition (HAR). However, they frequently encounter difficulties in accurately discerning intricate human actions in real-time situations. This study aims to fill a significant research gap by incorporating the Convolutional Block Attention Module (CBAM) into CNN architectures. The goal is to improve the extraction of features from video sequences. The CBAM boosts the performance of the network by selectively prioritizing significant spatial and channel-wise data, resulting in improved detection of subtle activity patterns and increased stability in categorization. CBAM's attention mechanism directly focuses and amplifies essential characteristics, which sets it apart from typical CNNs that lack a refined focus mechanism. This unique approach results in improved performance in behavior identification tests. The proposed CBAMenhanced model has been extensively tested on benchmark datasets, yielding an accuracy of 94.23% on the HMDB51 dataset. It also achieved competitive results of 83.4% and 88.9% on the UCF-101 and UCF-50 datasets, respectively. However, there is still a lack of study in comprehending how CBAM adjusts to different CNN architectures and its suitability in varied HAR situations beyond controlled datasets. In future studies, it is imperative for researchers to investigate the integration of CBAM with other CNN frameworks, assess its efficacy in practical scenarios, and explore multi-modal sensor fusion techniques to enhance its reliability and utility. This study showcases the ability of CBAM to enhance HAR capabilities and also paves the way for future research to improve activity identification systems for wider and more practical uses.
引用
收藏
页数:24
相关论文
共 50 条
  • [21] Microexpression Recognition Method Based on ADP-DSTN Feature Fusion and Convolutional Block Attention Module
    Song, Junfang
    Lei, Shanzhong
    Wu, Wenzhe
    ELECTRONICS, 2024, 13 (20)
  • [22] RT-CBAM: Refined Transformer Combined with Convolutional Block Attention Module for Underwater Image Restoration
    Ye, Renchuan
    Qian, Yuqiang
    Huang, Xinming
    SENSORS, 2024, 24 (18)
  • [23] Visual question answering on blood smear images using convolutional block attention module powered object detection
    Lubna, A.
    Kalady, Saidalavi
    Lijiya, A.
    VISUAL COMPUTER, 2025, 41 (01) : 739 - 757
  • [24] Fine-grained recognition of grape leaf diseases based on transfer learning and convolutional block attention module
    Wu, Canghai
    Gu, Xingxiang
    Xiong, Huanliang
    Huang, Huixin
    APPLIED SOFT COMPUTING, 2025, 172
  • [25] Convolutional Neural Network with Attention Module for Identification of Tunnel Seepage
    Chen, Qian
    Xiong, Chuanguo
    Lv, Weishan
    Shen, Ben
    Zeng, Baoshan
    Li, Jinming
    Feng, Chenzefang
    Hu, Zhou
    Zhu, Fulong
    TRANSPORTATION RESEARCH RECORD, 2022, 2676 (11) : 112 - 123
  • [26] A New Cyclic Spatial Attention Module for Convolutional Neural Networks
    Li Daihui
    Zeng Shangyou
    Li Wenhui
    Yang Lei
    2019 IEEE 11TH INTERNATIONAL CONFERENCE ON COMMUNICATION SOFTWARE AND NETWORKS (ICCSN 2019), 2019, : 607 - 611
  • [27] Deep Context Model (DCM): dual context-attention aware model for recognizing the heterogeneous human activities using smartphone sensors
    Kumar, Prabhat
    Suresh, S.
    EVOLVING SYSTEMS, 2024, 15 (04) : 1475 - 1486
  • [28] Efficient Spatial-Attention Module for Human Pose Estimation
    Tran, Tien-Dat
    Vo, Xuan-Thuy
    Nguyen, Duy-Linh
    Jo, Kang-Hyun
    FRONTIERS OF COMPUTER VISION, IW-FCV 2021, 2021, 1405 : 242 - 250
  • [29] Deep Learning for Recognizing Human Activities Using Motions of Skeletal Joints
    Phyo, Cho Nilar
    Zin, Thi Thi
    Tin, Pyke
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2019, 65 (02) : 243 - 252
  • [30] A Compound Eigenspace for Recognizing Directed Human Activities
    Diaf, Abdunnaser
    Boufama, Boubakeur
    Benlamri, Rachid
    IMAGE ANALYSIS AND RECOGNITION, PT II, 2012, 7325 : 122 - 129