A sequence models-based real-time multi-person action recognition method with monocular vision

被引:0
作者
Aolei Yang
Wei Lu
Wasif Naeem
Ling Chen
Minrui Fei
机构
[1] Shanghai University,School of Mechatronic Engineering and Automation
[2] Queen’s University Belfast,School of Electronics, Electrical Engineering and Computer Science
[3] Hunan Normal University,School of Engineering and Design
来源
Journal of Ambient Intelligence and Humanized Computing | 2023年 / 14卷
关键词
Action recognition; Human body skeleton; Feature construction; Sequence models; Computer vision;
D O I
暂无
中图分类号
学科分类号
摘要
In intelligent video surveillance under complex scenes, it is vital to identify the current actions of multi-target human bodies accurately and in real time. In this paper, a real-time multi-person action recognition method with monocular vision is proposed based on sequence models. Firstly, the key points of multi-target human body skeleton in the video are extracted by using the OpenPose algorithm. Then, the human action features are constructed, including limb direction vector and the skeleton height-width ratio. The multi-target human bodies tracking is then achieved by using the tracking algorithm. Next, the tracking results are matched with the action features, and the action recognition model is constructed, which includes the spatial branch based on Deep neural networks and the temporal branch based on Bi-directional RNN and Bi-directional long short-term memory networks. After pre-training, the model can be used to recognize the human body action from action features, and a recognition stabilizer is designed to minimize false alarms. Finally, extensive evaluations on the JHMDB dataset validate the effectiveness and the superiority of the proposed approach.
引用
收藏
页码:1877 / 1887
页数:10
相关论文
共 42 条
  • [1] Du W(2017)Recurrent spatial-temporal attention network for action recognition in videos IEEE Trans Image Process 27 1347-1360
  • [2] Wang Y(2013)EFM: evolutionary fuzzy model for dynamic activities recognition using a smartphone accelerometer Appl Intell 39 475-488
  • [3] Qiao Y(2013)A survey on human activity recognition using wearable sensors IEEE Commun Surv Tutor 15 1192-1209
  • [4] Fahim M(2017)Skeleton-based action recognition using spatio-temporal LSTM network with trust gates IEEE Trans Pattern Anal Mach Intell 40 3007-3021
  • [5] Fatima I(2017)Enhanced skeleton visualization for view invariant human action recognition Pattern Recogn 68 346-362
  • [6] Lara D(2020)Multilayer deep features with multiple kernel learning for action recognition Neurocomputing 399 65-74
  • [7] Oscar J(2014)Fusion of smartphone motion sensors for physical activity recognition Sensors 14 10146-10176
  • [8] Labrador A(2019)Temporal-spatial mapping for action recognition IEEE Trans Circuits Syst Video Technol 30 748-759
  • [9] Liu D(2018)Multi-stream CNN: learning representations based on human-related regions for action recognition Pattern Recogn 79 32-43
  • [10] Shahroudy M(2013)Dense trajectories and motion boundary descriptors for action recognition Int J Comput Vision 103 60-79