Multi-modal human aggression detection

被引:44
作者
Kooij, J. F. P. [1 ]
Liem, M. C. [1 ]
Krijnders, J. D. [2 ,3 ]
Andringa, T. C. [2 ]
Gavrila, D. M. [1 ]
机构
[1] Univ Amsterdam, Fac Sci, Intelligent Syst Lab, Amsterdam, Netherlands
[2] Univ Groningen, Artificial Intelligence, Auditory Cognit Grp, Groningen, Netherlands
[3] INCAS3, Cognit Syst Grp, Assen, Netherlands
关键词
Automated video surveillance; Multi-modal sensor fusion; Aggression detection; Dynamic Bayesian Network; EVENT RECOGNITION; MARKOV-MODELS; PEOPLE; TRACKING; LISTENERS; SPACE;
D O I
10.1016/j.cviu.2015.06.009
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a smart surveillance system named CASSANDRA, aimed at detecting instances of aggressive human behavior in public environments. A distinguishing aspect of CASSANDRA is the exploitation of complementary audio and video cues to disambiguate scene activity in real-life environments. From the video side, the system uses overlapping cameras to track persons in 3D and to extract features regarding the limb motion relative to the torso. From the audio side, it classifies instances of speech, screaming, singing, and kicking-object. The audio and video cues are fused with contextual cues (interaction, auxiliary objects); a Dynamic Bayesian Network (DBN) produces an estimate of the ambient aggression level. Our prototype system is validated on a realistic set of scenarios performed by professional actors at an actual train station to ensure a realistic audio and video noise setting. (C) 2015 Elsevier Inc. All rights reserved.
引用
收藏
页码:106 / 120
页数:15
相关论文
共 50 条
  • [31] Multi-Modal Ship Target Image Smoothing Based on Adaptive Mean Shift
    Liu, Zhaoying
    Bai, Xiangzhi
    Sun, Changming
    Zhou, Fugen
    Li, Yujian
    IEEE ACCESS, 2018, 6 : 12573 - 12586
  • [32] From the lab into the wild: Design and deployment methods for multi-modal tracking platforms
    Sommer, Philipp
    Kusy, Branislav
    Jurdak, Raja
    Kottege, Navinda
    Liu, Jiajun
    Zhao, Kun
    McKeown, Adam
    Westcott, David
    PERVASIVE AND MOBILE COMPUTING, 2016, 30 : 1 - 17
  • [33] A multi-modal evaluation of experimental pain and psychological function in women with carpometacarpal osteoarthritis
    Diaz, Tamara Ordonez
    Vasilopoulos, Terrie
    Wright, Thomas W.
    Cruz-Almeida, Yenisel
    Nichols, Jennifer A.
    OSTEOARTHRITIS AND CARTILAGE OPEN, 2024, 6 (04):
  • [34] A multi-modal probe for trace water and simultaneously monitoring lipid droplets and lysosomes
    Ye, Peng
    Zhang, Haitao
    Qu, Jianbo
    Zhu, Xiuzhong
    Hu, Qingfei
    Ma, Shanghong
    ORGANIC & BIOMOLECULAR CHEMISTRY, 2022, 20 (46) : 9184 - 9191
  • [35] Using multi-modal 3D contours and their relations for vision and robotics
    Baseski, Emre
    Pugeault, Nicolas
    Kalkan, Sinan
    Bodenhagen, Leon
    Piater, Justus H.
    Kruger, Norbert
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2010, 21 (08) : 850 - 864
  • [36] Capturing network properties with a functional form for the multi-modal macroscopic fundamental diagram
    Loder, Allister
    Dakic, Igor
    Bressan, Lea
    Ambuehl, Lukas
    Bliemer, Michiel C. J.
    Menendez, Monica
    Axhausen, Kay W.
    TRANSPORTATION RESEARCH PART B-METHODOLOGICAL, 2019, 129 : 1 - 19
  • [37] Multi-modal video event recognition based on association rules and decision fusion
    Guder, Mennan
    Cicekli, Nihan Kesim
    MULTIMEDIA SYSTEMS, 2018, 24 (01) : 55 - 72
  • [38] A randomized, controlled trial of a multi-modal exercise intervention in Huntington's disease
    Quinn, Lori
    Hamana, Katy
    Kelson, Mark
    Dawes, Helen
    Collett, Johnny
    Townson, Julia
    Roos, Raymund
    van der Plas, Anton Adriaan
    Reilmann, Ralf
    Frich, Jan C.
    Rickards, Hugh
    Rosser, Anne
    Busse, Monica
    PARKINSONISM & RELATED DISORDERS, 2016, 31 : 46 - 52
  • [39] Multi-modal video event recognition based on association rules and decision fusion
    Mennan Güder
    Nihan Kesim Çiçekli
    Multimedia Systems, 2018, 24 : 55 - 72
  • [40] UAV target following in complex occluded environments with adaptive multi-modal fusion
    Xu, Lele
    Wang, Teng
    Cai, Wenzhe
    Sun, Changyin
    APPLIED INTELLIGENCE, 2023, 53 (13) : 16998 - 17014