Self-supervised memory-guided and attention feature fusion for video anomaly detection

被引:0
|
作者
Jiang, Zitai [1 ]
Wang, Chuanxu [1 ]
Li, Jiajiong [1 ]
Zhao, Min [1 ]
Yang, Qingyang [1 ]
机构
[1] Qingdao Univ Sci & Technol, Dept Informat Sci & Technol, Zhonghan St, Qingdao, Peoples R China
基金
中国国家自然科学基金;
关键词
video anomaly detection; one-class classification; self-supervised; memory-guided; attention feature fusion;
D O I
10.1117/1.JEI.33.6.063035
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Video anomaly detection (VAD) aims to identify events and behaviors in video sequences that deviate from established normal patterns. Traditionally, unsupervised VAD has been viewed as the one-class classification (OCC) task, which is predicated on the use of training data that encompass only normal events and exclude any anomaly samples. This approach can recognize previously unseen anomaly patterns. However, it may misclassify unfamiliar normal patterns as anomalies. In addition, as anomalous samples are not included in OCC training, it leads to unclear classification boundaries, which reduces the generalization ability of the model. To address these challenges, we introduce the self-supervised memory-guided and attention feature fusion method, which models normal events using optimized memory modules and attention feature fusion modules. The method not only generates pseudo-normal and pseudo-anomaly data but also significantly enhances the capability of the model in key feature identification and exploitation, thereby improving the capture and generalization of complex data relationships. Experimental results from three benchmark datasets-UCSD Ped2, CUHK Avenue, and ShanghaiTech show that our method achieves AUROCs of 99.5%, 90.9%, and 81.8%, respectively, demonstrating the efficacy of our approach. Our code is available at https://github.com/jzt-dongli/Self-Supervised-Memory-guided-and-Attention-Feature-Fusion-for-Video-Anomaly-Detection. (c) 2024 SPIE and IS&T
引用
收藏
页数:24
相关论文
共 50 条
  • [1] Video anomaly detection with memory-guided multilevel embedding
    Zhou, Liuping
    Yang, Jing
    INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2023, 12 (01)
  • [2] Video anomaly detection with memory-guided multilevel embedding
    Liuping Zhou
    Jing Yang
    International Journal of Multimedia Information Retrieval, 2023, 12
  • [3] Memory-guided representation matching for unsupervised video anomaly detection
    Tao, Yiran
    Hu, Yaosi
    Chen, Zhenzhong
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 101
  • [4] Self-supervised Sparse Representation for Video Anomaly Detection
    Wu, Jhih-Ciang
    Hsieh, He-Yen
    Chen, Ding-Jie
    Fuh, Chiou-Shann
    Liu, Tyng-Luh
    COMPUTER VISION, ECCV 2022, PT XIII, 2022, 13673 : 729 - 745
  • [5] Innovative Video Anomaly Detection: TCN-AnoDetect With Self-Supervised Feature Learning
    Chiranjeevi, V. Rahul
    Malathi, D.
    TRANSACTIONS ON EMERGING TELECOMMUNICATIONS TECHNOLOGIES, 2025, 36 (01):
  • [6] OBJECT-CENTRIC AND MEMORY-GUIDED NORMALITY RECONSTRUCTION FOR VIDEO ANOMALY DETECTION
    Bergaoui, Khalil
    Naji, Yassine
    Setkov, Aleksandr
    Loesch, Angelique
    Gouiffes, Michele
    Audigier, Romaric
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 2691 - 2695
  • [7] Multimodal and multiscale feature fusion for weakly supervised video anomaly detection
    Sun, Wenwen
    Cao, Lin
    Guo, Yanan
    Du, Kangning
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [8] Self-Supervised Autoencoders for Visual Anomaly Detection
    Bauer, Alexander
    Nakajima, Shinichi
    Mueller, Klaus-Robert
    MATHEMATICS, 2024, 12 (24)
  • [9] Object-centric and memory-guided network-based normality modeling for video anomaly detection
    Chandrakala, S.
    Shalmiya, P.
    Srinivas, V
    Deepak, K.
    SIGNAL IMAGE AND VIDEO PROCESSING, 2022, 16 (07) : 2001 - 2007
  • [10] Learning Appearance-Motion Synergy via Memory-Guided Event Prediction for Video Anomaly Detection
    Guo, Chongye
    Wang, Hongbo
    Xia, Yingjie
    Feng, Guorui
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (03) : 1519 - 1531