Semisupervised Training of Deep Generative Models for High-Dimensional Anomaly Detection

被引:14
作者
Xie, Qin [1 ]
Zhang, Peng [2 ]
Yu, Boseon [2 ]
Choi, Jaesik [2 ]
机构
[1] Ulsan Natl Inst Sci & Technol, Dept Comp Sci & Engn, Ulsan 44919, South Korea
[2] Korea Adv Inst Sci & Technol, Grad Sch Artificial Intelligence, Daejeon 34141, South Korea
关键词
Anomaly detection; Data models; Semisupervised learning; Generative adversarial networks; Training; Generators; Unsupervised learning; deep generative models; semisupervised learning; variational autoencoder (VAE);
D O I
10.1109/TNNLS.2021.3095150
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Abnormal behaviors in industrial systems may be early warnings on critical events that may cause severe damages to facilities and security. Thus, it is important to detect abnormal behaviors accurately and timely. However, the anomaly detection problem is hard to solve in practice, mainly due to the rareness and the expensive cost to get the labels of the anomalies. Deep generative models parameterized by neural networks have achieved state-of-the-art performance in practice for many unsupervised and semisupervised learning tasks. We present a new deep generative model, Latent Enhanced regression/classification Deep Generative Model (LEDGM), for the anomaly detection problem with multidimensional data. Instead of using two-stage decoupled models, we adopt an end-to-end learning paradigm. Instead of conditioning the latent on the class label, LEDGM conditions the label prediction on the learned latent so that the optimization goal is more in favor of better anomaly detection than better reconstruction that the previously proposed deep generative models have been trained for. Experimental results on several synthetic and real-world small- and large-scale datasets demonstrate that LEDGM can achieve improved anomaly detection performance on multidimensional data with very sparse labels. The results also suggest that both labeled anomalies and labeled normal are valuable for semisupervised learning. Generally, our results show that better performance can be achieved with more labeled data. The ablation experiments show that both the original input and the learned latent provide meaningful information for LEDGM to achieve high performance.
引用
收藏
页码:2444 / 2453
页数:10
相关论文
共 36 条
[1]   GANomaly: Semi-supervised Anomaly Detection via Adversarial Training [J].
Akcay, Samet ;
Atapour-Abarghouei, Amir ;
Breckon, Toby P. .
COMPUTER VISION - ACCV 2018, PT III, 2019, 11363 :622-637
[2]  
An J., 2015, SPEC LECT IE
[3]  
Bin Z, 2019, PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P4433
[4]  
Cao, 2020, ARXIV200906847
[5]   Combining GANs and AutoEncoders for efficient anomaly detection [J].
Carrara, Fabio ;
Amato, Giuseppe ;
Brombin, Luca ;
Falchi, Fabrizio ;
Gennaro, Claudio .
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, :3939-3946
[6]  
Chalapathy R., 2019, ARXIV190103407
[7]  
Chapelle O., 2006, SEMISUPERVISED LEARN, V20, DOI DOI 10.1109/TNN.2009.2015974
[8]  
Donahue J., 2017, INT C LEARNING REPRE
[9]  
Dua D., 2017, UCI Machine Learning Repository
[10]   A Dataset to Support Research in the Design of Secure Water Treatment Systems [J].
Goh, Jonathan ;
Adepu, Sridhar ;
Junejo, Khurum Nazir ;
Mathur, Aditya .
CRITICAL INFORMATION INFRASTRUCTURES SECURITY (CRITIS 2016), 2018, 10242 :88-99