Recurrent Neural Network Attention Mechanisms for Interpretable System Log Anomaly Detection

被引:106
作者
Brown, Andy [1 ]
Tuor, Aaron [2 ]
Hutchinson, Brian [1 ,3 ,4 ]
Nichols, Nicole [2 ]
机构
[1] Western Washington Univ, Bellingham, WA 98225 USA
[2] Pacific Northwest Natl Lab, Richland, WA 99352 USA
[3] PNNL, Richland, WA USA
[4] WWU, Bellingham, WA USA
来源
PROCEEDINGS OF THE 1ST WORKSHOP ON MACHINE LEARNING FOR COMPUTING SYSTEMS (MLCS 2018) | 2018年
关键词
Anomaly detection; Attention; Recurrent Neural Networks; Interpretable Machine Learning; Online Training; System Log Analysis;
D O I
10.1145/3217871.3217872
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning has recently demonstrated state-of-the art performance on key tasks related to the maintenance of computer systems, such as intrusion detection, denial of service attack detection, hardware and software system failures, and malware detection. In these contexts, model interpretability is vital for administrator and analyst to trust and act on the automated analysis of machine learning models. Deep learning methods have been criticized as black box oracles which allow limited insight into decision factors. In this work we seek to bridge the gap between the impressive performance of deep learning models and the need for interpretable model introspection. To this end we present recurrent neural network (RNN) language models augmented with attention for anomaly detection in system logs. Our methods are generally applicable to any computer system and logging source. By incorporating attention variants into our RNN language models we create opportunities for model introspection and analysis without sacrificing state-of-the art performance. We demonstrate model performance and illustrate model interpretability on an intrusion detection task using the Los Alamos National Laboratory (LANL) cyber security dataset, reporting upward of 0.99 area under the receiver operator characteristic curve despite being trained only on a single day's worth of data.
引用
收藏
页数:8
相关论文
共 23 条
[1]  
[Anonymous], P NIPS
[2]  
[Anonymous], 2016, ARXIV161001644
[3]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[4]  
Chang Chun-Hao., 2017, Proc. NIPS
[5]  
Daniluk Michal, 2017, 5 INT C LEARNING REP
[6]  
Du Min, 2017, P SIGSAC
[7]  
Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.8.1735, 10.1007/978-3-642-24797-2, 10.1162/neco.1997.9.1.1]
[8]  
Ke Zhang, 2016, P IEEE BIG DAT
[9]  
Kent AD., 2016, DYNAMIC NETWORKS CYB, P37, DOI [DOI 10.1142/9781786340757_0002, DOI 10.1142/97817863407570002]
[10]  
Kingma D. P., P 3 INT C LEARN REPR