A Comparative Study on Deep Learning and Machine Learning Models for Human Action Recognition in Aerial Videos

被引:4
作者
Kapoor, Surbhi [1 ]
Sharma, Akashdeep [1 ]
Verma, Amandeep [1 ]
Dhull, Vishal [1 ]
Goyal, Chahat [1 ]
机构
[1] Panjab Univ, UIET, Chandigarh, India
关键词
Human action recognition; openpose; unmanned aerial vehicle; kNN; decision trees; random forests; SVM; multilayer perceptron; LSTM;
D O I
10.34028/iajit/20/4/2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unmanned Aerial Vehicle)UAV( finds its significant application in video surveillance due to its low cost, high portability and fast-mobility. In this paper, the proposed approach focuses on recognizing the human activity in aerial video sequences through various keypoints detected on the human body via OpenPose. The detected keypoints are passed onto machine learning and deep learning classifiers for classifying the human actions. Experimental results demonstrate that multilayer perceptron and SVM outperformed all the other classifiers by reporting an accuracy of 87.80% and 87.77% respectively whereas LSTM did not produce very good results as compared to other classifiers. Stacked Long Short-Term Memory networks (LSTM( produced an accuracy of 71.30% and Bidirectional LSTM yielded an accuracy of 76.04%. The results also indicate that machine learning models performed better than deep learning models. The major reason for this finding is the lesser availability of data and the deep learning models being data hungry models require a large amount of data to work upon. The paper also analyses the failure cases of OpenPose by testing the system on aerial videos captured by a drone flying at a higher altitude. This work provides a baseline for validating machine learning classifiers and deep learning classifiers against recognition of human action from aerial videos.
引用
收藏
页码:567 / 574
页数:8
相关论文
共 16 条
  • [1] Vision Based Victim Detection from Unmanned Aerial Vehicles
    Andriluka, Mykhaylo
    Schnitzspan, Paul
    Meyer, Johannes
    Kohlbrecher, Stefan
    Petersen, Karen
    von Stryk, Oskar
    Roth, Stefan
    Schiele, Bernt
    [J]. IEEE/RSJ 2010 INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2010), 2010, : 1740 - 1747
  • [2] Bolin J, 2017, AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, P1484
  • [3] OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields
    Cao, Zhe
    Hidalgo, Gines
    Simon, Tomas
    Wei, Shih-En
    Sheikh, Yaser
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (01) : 172 - 186
  • [4] Intelligent Human-UAV Interaction System with Joint Cross-Validation over Action-Gesture Recognition and Scene Understanding
    Chen, Bo
    Hua, Chunsheng
    Li, Decai
    He, Yuqing
    Han, Jianda
    [J]. APPLIED SCIENCES-BASEL, 2019, 9 (16):
  • [5] Li MP, 2018, INT C PATT RECOG, P115, DOI 10.1109/ICPR.2018.8546194
  • [6] Penmetsa S., 2014, Electron. Lett. Comput. Vis. Image Anal, V13, P18
  • [7] Perera A., 2018, P EUR C COMP VIS WOR
  • [8] Drone-Action: An Outdoor Recorded Drone Video Dataset for Action Recognition
    Perera, Asanka G.
    Law, Yee Wei
    Chahl, Javaan
    [J]. DRONES, 2019, 3 (04) : 1 - 16
  • [9] Human Pose and Path Estimation from Aerial Video Using Dynamic Classifier Selection
    Perera, Asanka G.
    Law, Yee Wei
    Chahl, Javaan
    [J]. COGNITIVE COMPUTATION, 2018, 10 (06) : 1019 - 1041
  • [10] DeepCut: Joint Subset Partition and Labeling for Multi Person Pose Estimation
    Pishchulin, Leonid
    Insafutdinov, Eldar
    Tang, Siyu
    Andres, Bjoern
    Andriluka, Mykhaylo
    Gehler, Peter
    Schiele, Bernt
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 4929 - 4937