Decouple and Resolve: Transformer-Based Models for Online Anomaly Detection From Weakly Labeled Videos

被引:24
作者
Liu, Tianshan [1 ]
Zhang, Cong [1 ]
Lam, Kin-Man [1 ,2 ]
Kong, Jun [3 ]
机构
[1] Hong Kong Polytech Univ, Dept Elect & Informat Engn, Hong Kong, Peoples R China
[2] Ctr Adv Reliabil & Safety, Hong Kong, Peoples R China
[3] Jiangnan Univ, Minist Educ, Key Lab Adv Proc Control Light Ind, Wuxi 214122, Peoples R China
关键词
Videos; Task analysis; Transformers; Anomaly detection; Proposals; Training; Annotations; Online video anomaly detection; weakly supervised learning; multi-task learning; long-short-term context;
D O I
10.1109/TIFS.2022.3216479
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
As one of the vital topics in intelligent surveillance, weakly supervised online video anomaly detection (WS-OVAD) aims to identify the ongoing anomalous events moment-to-moment in streaming videos, trained with only video-level annotations. Previous studies tended to utilize a unified single-stage framework, which struggled to simultaneously address the issues of online constraints and weakly supervised settings. To solve this dilemma, in this paper, we propose a two-stage-based framework, namely "decouple and resolve " (DAR), which consists of two modules, i.e., temporal proposal producer (TPP) and online anomaly localizer (OAL). With the supervision of video-level binary labels, the TPP module targets fully exploiting hierarchical temporal relations among snippets for generating precise snippet-level pseudo-labels. Then, given fine-grained supervisory signals produced by TPP, the Transformer-based OAL module is trained to aggregate both the useful cues retrieved from historical observations and anticipated future semantics, for making predictions at the current time step. Both the TPP and OAL modules are jointly trained to share the beneficial knowledge in a multi-task learning paradigm. Extensive experimental results on three public data sets validate the superior performance of the proposed DAR framework over the competing methods.
引用
收藏
页码:15 / 28
页数:14
相关论文
共 57 条
[1]   ViViT: A Video Vision Transformer [J].
Arnab, Anurag ;
Dehghani, Mostafa ;
Heigold, Georg ;
Sun, Chen ;
Lucic, Mario ;
Schmid, Cordelia .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :6816-6826
[2]  
Bai SJ, 2018, Arxiv, DOI [arXiv:1803.01271, DOI 10.48550/ARXIV.1803.01271]
[3]  
Bertasius G, 2021, Arxiv, DOI [arXiv:2102.05095, DOI 10.48550/ARXIV.2102.05095]
[4]   Multiple instance learning: A survey of problem characteristics and applications [J].
Carbonneau, Marc-Andre ;
Cheplygina, Veronika ;
Granger, Eric ;
Gagnon, Ghyslain .
PATTERN RECOGNITION, 2018, 77 :329-353
[5]   End-to-End Object Detection with Transformers [J].
Carion, Nicolas ;
Massa, Francisco ;
Synnaeve, Gabriel ;
Usunier, Nicolas ;
Kirillov, Alexander ;
Zagoruyko, Sergey .
COMPUTER VISION - ECCV 2020, PT I, 2020, 12346 :213-229
[6]   Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset [J].
Carreira, Joao ;
Zisserman, Andrew .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :4724-4733
[7]   Contrastive Attention for Video Anomaly Detection [J].
Chang, Shuning ;
Li, Yanchao ;
Shen, Shengmei ;
Feng, Jiashi ;
Zhou, Zhiying .
IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 :4067-4076
[8]   Modeling temporal structure with LSTM for online action detection [J].
De Geest, Roeland ;
Tuytelaars, Tinne .
2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, :1549-1557
[9]   Online Action Detection [J].
De Geest, Roeland ;
Gavves, Efstratios ;
Ghodrati, Amir ;
Li, Zhenyang ;
Snoek, Cees ;
Tuytelaars, Tinne .
COMPUTER VISION - ECCV 2016, PT V, 2016, 9909 :269-284
[10]  
Dosovitskiy Alexey, 2021, P ICLR