共 50 条
Deep Learning for Automatic Violence Detection: Tests on the AIRTLab Dataset
被引:31
作者:
Sernani, Paolo
[1
]
Falcionelli, Nicola
[1
]
Tomassini, Selene
[1
]
Contardo, Paolo
[1
,2
]
Dragoni, Aldo Franco
[1
]
机构:
[1] Univ Politecn Marche, Dipartimento Ingn Informaz, I-60131 Ancona, Italy
[2] Gabinetto Interreg Polizia Sci Marche & Abruzzo, I-60129 Ancona, Italy
来源:
关键词:
Atmospheric modeling;
Sports;
Three-dimensional displays;
Feature extraction;
Solid modeling;
Task analysis;
Deep learning;
Convolutional long short-term memory;
convolutional neural network;
deep learning;
support vector machine;
violence detection;
D O I:
10.1109/ACCESS.2021.3131315
中图分类号:
TP [自动化技术、计算机技术];
学科分类号:
0812 ;
摘要:
Following the growing availability of video surveillance cameras and the need for techniques to automatically identify events in video footages, there is an increasing interest towards automatic violence detection in videos. Deep learning-based architectures, such as 3D Convolutional Neural Networks, demonstrated their capability of extracting spatio-temporal features from videos, being effective in violence detection. However, friendly behaviours or fast moves such as hugs, small hits, claps, high fives, etc., can still cause false positives, interpreting a harmless action as violent. To this end, we present three deep learning-based models for violence detection and test them on the AIRTLab dataset, a novel dataset designed to check the robustness of algorithms against false positives. The objective is twofold: on one hand, we compute accuracy metrics on the three proposed models (two are based on transfer learning and one is trained from scratch), building a baseline of metrics for the AIRTLab dataset; on the other hand, we validate the capability of the proposed dataset of challenging the robustness to false positives. The results of the proposed models are in line with the scientific literature, in terms of accuracy, with transfer learning-based networks exhibiting better generalization capabilities than the trained from scratch network. Moreover, the tests highlighted that most of the classification errors concern the identification of non-violent clips, validating the design of the proposed dataset. Finally, to demonstrate the significance of the proposed models, the paper presents a comparison with the related literature, as well as with models based on well-established pre-trained 2D Convolutional Neural Networks (2D CNNs). Such comparison highlights that 3D models get better accuracy performance than time distributed 2D CNNs (merged with a recurrent module) in processing the spatio-temporal features of video clips. The source code of the experiments and the AIRTLab dataset are available in public repositories.
引用
收藏
页码:160580 / 160595
页数:16
相关论文
共 50 条