A Deep Bidirectional LSTM Model Enhanced by Transfer-Learning-Based Feature Extraction for Dynamic Human Activity Recognition

被引:15
作者
Hassan, Najmul [1 ]
Miah, Abu Saleh Musa [1 ]
Shin, Jungpil [1 ]
机构
[1] Univ Aizu, Sch Comp Sci & Engn, Aizu Wakamatsu 9658580, Japan
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 02期
关键词
pre-trained neural networks; HAR; MobileNetv2; deep bidirectional LSTM; NEURAL-NETWORKS; DATA STREAMS; CNN; FUSION;
D O I
10.3390/app14020603
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Dynamic human activity recognition (HAR) is a domain of study that is currently receiving considerable attention within the fields of computer vision and pattern recognition. The growing need for artificial-intelligence (AI)-driven systems to evaluate human behaviour and bolster security underscores the timeliness of this research. Despite the strides made by numerous researchers in developing dynamic HAR frameworks utilizing diverse pre-trained architectures for feature extraction and classification, persisting challenges include suboptimal performance accuracy and the computational intricacies inherent in existing systems. These challenges arise due to the vast video-based datasets and the inherent similarity in the data. To address these challenges, we propose an innovative, dynamic HAR technique employing a deep-learning-based, deep bidirectional long short-term memory (Deep BiLSTM) model facilitated by a pre-trained transfer-learning-based feature-extraction approach. Our approach begins with the utilization of Convolutional Neural Network (CNN) models, specifically MobileNetV2, for extracting deep-level features from video frames. Subsequently, these features are fed into an optimized deep bidirectional long short-term memory (Deep BiLSTM) network to discern dependencies and process data, enabling optimal predictions. During the testing phase, an iterative fine-tuning procedure is introduced to update the high parameters of the trained model, ensuring adaptability to varying scenarios. The proposed model's efficacy was rigorously evaluated using three benchmark datasets, namely UCF11, UCF Sport, and JHMDB, achieving notable accuracies of 99.20%, 93.3%, and 76.30%, respectively. This high-performance accuracy substantiates the superiority of our proposed model, signaling a promising advancement in the domain of activity recognition.
引用
收藏
页数:18
相关论文
共 72 条
  • [11] Two Stream LSTM : A Deep Fusion Framework for Human Action Recognition
    Gammulle, Harshala
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    [J]. 2017 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2017), 2017, : 177 - 186
  • [12] A survey on deep learning techniques for image and video semantic segmentation
    Garcia-Garcia, Alberto
    Orts-Escolano, Sergio
    Oprea, Sergiu
    Villena-Martinez, Victor
    Martinez-Gonzalez, Pablo
    Garcia-Rodriguez, Jose
    [J]. APPLIED SOFT COMPUTING, 2018, 70 : 41 - 65
  • [13] Green Resource Allocation Based on Deep Reinforcement Learning in Content-Centric IoT
    He, Xiaoming
    Wang, Kun
    Huang, Huawei
    Miyazaki, Toshiaki
    Wang, Yixuan
    Guo, Song
    [J]. IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2020, 8 (03) : 781 - 796
  • [14] Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.8.1735, 10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
  • [15] Human Behavior Recognition from Multiview Videos
    Hsueh, Yu-Ling
    Lie, Wen-Nung
    Guo, Guan-You
    [J]. INFORMATION SCIENCES, 2020, 517 : 275 - 296
  • [16] Hybrid deep neural network model for human action recognition
    Ijjina, Earnest Paul
    Mohan, Chalavadi Krishna
    [J]. APPLIED SOFT COMPUTING, 2016, 46 : 936 - 952
  • [17] A new hybrid deep learning model for human action recognition
    Jaouedi, Neziha
    Boujnah, Noureddine
    Bouhlel, Salim
    [J]. JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2020, 32 (04) : 447 - 453
  • [18] Towards understanding action recognition
    Jhuang, Hueihan
    Gall, Juergen
    Zuffi, Silvia
    Schmid, Cordelia
    Black, Michael J.
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 3192 - 3199
  • [19] Cover the Violence: A Novel Deep-Learning-Based Approach Towards Violence-Detection in Movies
    Khan, Samee Ullah
    Ul Haq, Ijaz
    Rho, Seungmin
    Baik, Sung Wook
    Lee, Mi Young
    [J]. APPLIED SCIENCES-BASEL, 2019, 9 (22):
  • [20] Active and adaptive ensemble learning for online activity recognition from data streams
    Krawczyk, Bartosz
    [J]. KNOWLEDGE-BASED SYSTEMS, 2017, 138 : 69 - 78