TriFusion hybrid model for human activity recognition

被引:0
作者
Ahmed, Md Foysal [1 ]
He, Gang [1 ,2 ]
Wang, Sikai [1 ]
机构
[1] Southwest Univ Sci & Technol, Sch Comp Sci & Technol, Mianyang 621010, Sichuan, Peoples R China
[2] Mianyang Cent Hosp, NHC Key Lab Nucl Technol Med Transformat, Mianyang 621010, Sichuan, Peoples R China
关键词
Human activity recognition; TriFusion model; VGG16; BiGRU; Transfer learning;
D O I
10.1007/s11760-024-03487-5
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Human activity recognition (HAR) remains a challenging problem in computer vision due to the unpredictable nature of human activities. In recent years, researchers have proposed various hybrid models for HAR, focusing on spatial, temporal, or both aspects. The spatial and temporal feature-based models have limitations in fully capturing the features and exhibit low accuracy during the training and testing. Addressing these challenges, we present a TriFusion hybrid model that integrates spatial, temporal, and high-level features to enhance accuracy in HAR. In this paper, different from previous fusion methods, we propose a novel approach by combining three deep learning model architectures. Specifically, VGG16 spatial features are fed into BiGRU for temporal feature extraction and also directly connected to TriFusion, alongside the BiGRU output, while transfer learning (ResNet18) is also connected to TriFusion. Our TriFusion model achieved an average accuracy of 99.92% on the UCF101 dataset and 99.78% on the HMDB51 dataset, demonstrating its suitability for real-time deployment in HAR applications. Our TriFusion model, designed for HAR tasks, exhibits promising applications across various AI domains, including human-computer interaction and diverse classification tasks. The code of TriFusion is publicly accessible at https://github.com/TripleTheGreatDali/TriFusionHAR.
引用
收藏
页码:8469 / 8476
页数:8
相关论文
共 34 条
[1]   A Survey of Vision-Based Transfer Learning in Human Activity Recognition [J].
Adama, David Ada ;
Lotfi, Ahmad ;
Ranson, Robert .
ELECTRONICS, 2021, 10 (19)
[2]   Insights into Internet of Medical Things (IoMT): Data fusion, security issues and potential solutions [J].
Ahmed, Shams Forruque ;
Bin Alam, Md. Sakib ;
Afrin, Shaila ;
Rafa, Sabiha Jannat ;
Rafa, Nazifa ;
Gandomi, Amir H. .
INFORMATION FUSION, 2024, 102
[3]   Vision-based human activity recognition: a survey [J].
Beddiar, Djamila Romaissa ;
Nini, Brahim ;
Sabokrou, Mohammad ;
Hadid, Abdenour .
MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (41-42) :30509-30555
[4]  
Bilal Muhammad, 2020, 2020 International Conference on Computational Science and Computational Intelligence (CSCI), P701, DOI 10.1109/CSCI51800.2020.00129
[5]   An Overview of the Vision-Based Human Action Recognition Field [J].
Camarena, Fernando ;
Gonzalez-Mendoza, Miguel ;
Chang, Leonardo ;
Cuevas-Ascencio, Ricardo .
MATHEMATICAL AND COMPUTATIONAL APPLICATIONS, 2023, 28 (02)
[6]   A feature extraction and machine learning framework for bearing fault diagnosis [J].
Cui, Bodi ;
Weng, Yang ;
Zhang, Ning .
RENEWABLE ENERGY, 2022, 191 :987-997
[7]  
Dedeoglu Y, 2006, LECT NOTES COMPUT SC, V3979, P64
[8]   RETRACTED: AGTH-Net: Attention-Based Graph Convolution-Guided Third-Order Hourglass Network for Sports Video Classification (Retracted Article) [J].
Gao, Ming ;
Cai, Weiwei ;
Liu, Runmin .
JOURNAL OF HEALTHCARE ENGINEERING, 2021, 2021
[9]  
Garg A., 2022, COMPUTATIONS BEVERLI, P1, DOI DOI 10.2514/6.2022-1034
[10]   Subtransfer Learning in Human Activity Recognition: Boosting the Outlier User Accuracy [J].
Kanjilal, Ria ;
Kucuk, Muhammed F. ;
Uysal, Ismail .
IEEE SENSORS JOURNAL, 2023, 23 (20) :25005-25015