HARTIV: Human Activity Recognition Using Temporal Information in Videos

被引:4
作者
Deotale, Disha [1 ]
Verma, Madhushi [2 ]
Suresh, P. [3 ]
Jangir, Sunil Kumar [4 ]
Kaur, Manjit [2 ]
Idris, Sahar Ahmed [5 ]
Alshazly, Hammam [6 ]
机构
[1] SPPU Univ, GH Raisoni Inst Engn & Technol, CSE Dept, Pune, Maharashtra, India
[2] Bennett Univ, CSE Dept, Greater Noida, India
[3] TML Business Serv Ltd, Pune, Maharashtra, India
[4] Anand Int Coll Engn, CSE Dept, Jaipur, Rajasthan, India
[5] King Khalid Univ, Coll Ind Engn, Abha, Saudi Arabia
[6] South Valley Univ, Fac Comp & Informat, Qena 83523, Egypt
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2022年 / 70卷 / 02期
关键词
Action recognition; human activity recognition; untrimmed video; deep learning; convolutional neural networks;
D O I
10.32604/cmc.2022.020655
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Nowadays, the most challenging and important problem of computer vision is to detect human activities and recognize the same with temporal information from video data. The video datasets are generated using cameras available in various devices that can be in a static or dynamic position and are referred to as untrimmed videos. Smarter monitoring is a historical necessity in which commonly occurring, regular, and out-of-the-ordinary activities can be automatically identified using intelligence systems and computer vision technology. In a long video, human activity may be present anywhere in the video. There can be a single or multiple human activities present in such videos. This paper presents a deep learning-based methodology to identify the locally present human activities in the video sequences captured by a single wide-view camera in a sports environment. The recognition process is split into four parts: firstly, the video is divided into different set of frames, then the human body part in a sequence of frames is identified, next process is to identify the human activity using a convolutional neural network and finally the time information of the observed postures for each activity is determined with the help of a deep learning algorithm. The proposed approach has been tested on two different sports datasets including ActivityNet and THUMOS. Three sports activities like swimming, cricket bowling and high jump have been considered in this paper and classified with the temporal information i.e., the start and end time for every activity present in the video. The convolutional neural network and long short-term memory are used for feature extraction of temporal action recognition from video data of sports activity. The outcomes show that the proposed method for activity recognition in the sports domain outperforms the existing methods.
引用
收藏
页码:3919 / 3938
页数:20
相关论文
共 32 条
  • [1] Improving bag-of-poses with semi-temporal pose descriptors for skeleton-based action recognition
    Agahian, Saeid
    Negin, Farhood
    Kose, Cemal
    [J]. VISUAL COMPUTER, 2019, 35 (04) : 591 - 607
  • [2] [Anonymous], 2017, CoRR abs/1708.05038
  • [3] Temporal detection and analysis of guideline interactions
    Anselma, Luca
    Piovesan, Luca
    Terenziani, Paolo
    [J]. ARTIFICIAL INTELLIGENCE IN MEDICINE, 2017, 76 : 40 - 62
  • [4] Vision-based human activity recognition: a survey
    Beddiar, Djamila Romaissa
    Nini, Brahim
    Sabokrou, Mohammad
    Hadid, Abdenour
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (41-42) : 30509 - 30555
  • [5] Human Activity Recognition with Convolutional Neural Networks
    Bevilacqua, Antonio
    MacDonald, Kyle
    Rangarej, Aamina
    Widjaya, Venessa
    Caulfield, Brian
    Kechadi, Tahar
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2018, PT III, 2019, 11053 : 541 - 552
  • [6] Bhuyan M. K, 2019, COMPUTER VISION IMAG, V1, P312
  • [7] Boualia SN, 2019, INT WIREL COMMUN, P1468, DOI [10.1109/IWCMC.2019.8766694, 10.1109/iwcmc.2019.8766694]
  • [8] Heilbron FC, 2015, PROC CVPR IEEE, P961, DOI 10.1109/CVPR.2015.7298698
  • [9] Action recognition on continuous video
    Chang, Y. L.
    Chan, C. S.
    Remagnino, P.
    [J]. NEURAL COMPUTING & APPLICATIONS, 2021, 33 (04) : 1233 - 1243
  • [10] Deotale D., 2020, ICICNIS 2020, P596