HAREDNet: A deep learning based architecture for autonomous video surveillance by recognizing human actions

被引:36
作者
Nasir, Inzamam Mashood [1 ]
Raza, Mudassar [1 ]
Shah, Jamal Hussain [1 ]
Wang, Shui-Hua [2 ]
Tariq, Usman [3 ]
Khan, Muhammad Attique [4 ]
机构
[1] COMSATS Univ Islamabad, Dept Comp Sci, Wah Campus, Wah Cantt, Pakistan
[2] Univ Leicester, Dept Math, Leicester, Leics, England
[3] Prince Sattam Bin Abdulaziz Univ, Coll Comp Engn & Sci, Al Kharaj, Saudi Arabia
[4] HITEC Univ Taxila, Dept Comp Sci, Taxila, Pakistan
关键词
Deep Convolutional Neural Network; Human Action Recognition; Weighted fusion; CvQDA; Encoder-Decoder CNN architecture;
D O I
10.1016/j.compeleceng.2022.107805
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Human Action Recognition (HAR) is still considered as a significant research area due to its emerging real-time applications like video surveillance, automated surveillance, real-time tracking and resecue missions. HAR domain still have gaps to cover, i.e., random changes in human variations, clothes, illumination, and backgrounds. Different camera settings, viewpoints and inter-class similarities have increased the complexity of this domain. The above-mentioned challenges in uncontrolled environment have ultimately reduced the performances of many well-designed models. The primary objective of this research is to propose and design an automated recognition system by overcoming these afore-mentioned issues. Redundant features and excessive computational time for the training and prediction process has also been a noteworthy problem. In this article, a hybrid recognition technique called HAREDNet is proposed, which has a) Encoder-Decoder Network (EDNet) to extract deep features; b) improved Scale-Invariant Feature Transform (iSIFT), improved Gabor (iGabor) and Local Maximal Occurrence (LOMO) techniques to extract local features; c) Cross-view Quadratic Discriminant Analysis (CvQDA) algorithm to reduce the feature redundancy; and d) weighted fusion strategy to merge properties of different essential features. The proposed technique is evaluated on three (3) publicly available datasets, including NTU RGB+D, HMDB51, and UCF-101, and achieved average recognition accuracy of 97.45%, 80.58%, and 97.48%, respectively, which is better than previously proposed methods.
引用
收藏
页数:17
相关论文
共 29 条
[1]   Human action recognition using bag of global and local Zernike moment features [J].
Aly, Saleh ;
Sayed, Asmaa .
MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (17) :24923-24953
[2]  
Baisa N.L., 2021, ARXIV PREPRINT ARXIV
[3]   JOLO-GCN: Mining Joint-Centered Light-Weight Information for Skeleton-Based Action Recognition [J].
Cai, Jinmiao ;
Jiang, Nianjuan ;
Han, Xiaoguang ;
Jia, Kui ;
Lu, Jiangbo .
2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, :2734-2743
[4]   Deep network for human action recognition using Weber motion [J].
Chaudhary, Sachin ;
Murala, Subrahmanyam .
NEUROCOMPUTING, 2019, 367 :207-216
[5]   Spatial-temporal channel-wise attention network for action recognition [J].
Chen, Lin ;
Liu, Yungang ;
Man, Yongchao .
MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (14) :21789-21808
[6]  
Du Y, 2015, PROC CVPR IEEE, P1110, DOI 10.1109/CVPR.2015.7298714
[7]   Two Stream LSTM : A Deep Fusion Framework for Human Action Recognition [J].
Gammulle, Harshala ;
Denman, Simon ;
Sridharan, Sridha ;
Fookes, Clinton .
2017 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2017), 2017, :177-186
[8]   Adaptive Fusion and Category-Level Dictionary Learning Model for Multiview Human Action Recognition [J].
Gao, Zan ;
Xuan, Hai-Zhen ;
Zhang, Hua ;
Wan, Shaohua ;
Choo, Kim-Kwang Raymond .
IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (06) :9280-9293
[9]   CGA: a new feature selection model for visual human action recognition [J].
Guha, Ritam ;
Khan, Ali Hussain ;
Singh, Pawan Kumar ;
Sarkar, Ram ;
Bhattacharjee, Debotosh .
NEURAL COMPUTING & APPLICATIONS, 2021, 33 (10) :5267-5286
[10]  
Huang LL, 2004, SIXTH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, PROCEEDINGS, P397