Enhancing Human Activity Detection and Classification Using Fine Tuned Attention-Based Transformer Models

被引:0
|
作者
Ram Kumar Yadav [1 ]
A. Daniel [1 ]
Vijay Bhaskar Semwal [2 ]
机构
[1] Amity University,Department of Computer Science and Engineering
[2] MANIT,Department of Computer Science and Engineering
关键词
Machine learning; Deep learning; HAR; Data augmentation; Attention-based transformers;
D O I
10.1007/s42979-024-03445-5
中图分类号
学科分类号
摘要
Recognition of human activity is an active research area. It uses the Internet of Things, Sensory methods, Machine Learning, and Deep Learning techniques to assist various application fields like home monitoring, robotics, surveillance, and healthcare. However, researchers face problems such as time complexity, more execution time of the model, and classification accuracy. This paper introduces a novel approach to overcome the issue as mentioned earlier by using the deep learning transformer model such as ViT(Vision Transformer), DieT(Data-efficient image Transformers), and SwinV2 transformer, which are used for image-based datasets (i.e., Standard40, MPII human pose) and VideoMAE transformer is used for video-based UCF101 and HMDB51 datasets. The approaches achieved remarkable accuracy in classifying human activities. Evaluations using the ViT, DeiT, and Swin transformer V2 with Stanford40 are 90.8%, 90.7%, and 88%; similarly, MPII Human Pose datasets show 87%, 85.6%, and 87.1%. In addition, this paper's method has achieved remarkable accuracies of 94.15% and 78.44%, respectively, when applying the VideoMAE transformer to video-based activity recognition on the UCF101 and HMDB51 datasets. These findings emphasize the efficacy of the attention-based transformer (i.e., ViT, DeiT, SwinV2, and VideoMAE) model and the novelty of earlier no-result evaluation on these various datasets with attention-based transformers.
引用
收藏
相关论文
共 50 条
  • [21] Robust Vulnerability Detection in Solidity-Based Ethereum Smart Contracts Using Fine-Tuned Transformer Encoder Models
    Le, Thi-Thu-Huong
    Kim, Jaehyun
    Lee, Sangmyeong
    Kim, Howon
    IEEE ACCESS, 2024, 12 : 154700 - 154717
  • [22] Understanding stance classification of BERT models: an attention-based framework
    Carlos Abel Córdova Sáenz
    Karin Becker
    Knowledge and Information Systems, 2024, 66 : 419 - 451
  • [23] Speech Emotion Classification Using Attention-Based LSTM
    Xie, Yue
    Liang, Ruiyu
    Liang, Zhenlin
    Huang, Chengwei
    Zou, Cairong
    Schuller, Bjoern
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2019, 27 (11) : 1675 - 1685
  • [24] A Hybrid Attention-Based Transformer Model for Arabic News Classification Using Text Embedding and Deep Learning
    Hossain, Md. Mithun
    Hossain, Md. Shakil
    Safran, Mejdl
    Alfarhood, Sultan
    Alfarhood, Meshal
    F. Mridha, M.
    IEEE ACCESS, 2024, 12 : 198046 - 198066
  • [25] An Attention-Based Approach to Enhance the Detection and Classification of Android Malware
    Ghourabi, Abdallah
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 80 (02): : 2743 - 2760
  • [26] Three-classification face manipulation detection using attention-based feature decomposition
    Cao, Yungui
    Chen, Jiazhen
    Huang, Liqing
    Huang, Tianqian
    Ye, Feng
    COMPUTERS & SECURITY, 2023, 125
  • [27] Enhancing breast cancer histopathological image classification using attention-based high order covariance pooling
    Waqas, Muhammad
    Ahmed, Amr
    Maul, Tomas
    Liao, Iman Yi
    Neural Computing and Applications, 2024, 36 (36) : 23275 - 23293
  • [28] Machine Translation Using Improved Attention-based Transformer with Hybrid Input
    Abrishami, Mahsa
    Rashti, Mohammad Javad
    Naderan, Marjan
    2020 6TH INTERNATIONAL CONFERENCE ON WEB RESEARCH (ICWR), 2020, : 52 - 57
  • [29] Attention-based supervised contrastive learning on fine-grained image classification
    Li, Qian
    Wu, Weining
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (03)
  • [30] Consistent Video Inpainting Using Axial Attention-Based Style Transformer
    Junayed, Masum Shah
    Islam, Md Baharul
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 7494 - 7504