Security Analysis of Online Centroid Anomaly Detection

被引:0
作者
Kloft, Marius [1 ,3 ]
Laskov, Pavel [2 ]
机构
[1] Tech Univ Berlin, Machine Learing Grp, D-10587 Berlin, Germany
[2] Univ Tubingen, Wilhelm Schickard Inst Comp Sci, D-72076 Tubingen, Germany
[3] Korea Univ, Dept Brain & Cognit Engn, Seoul 136713, South Korea
基金
新加坡国家研究基金会;
关键词
anomaly detection; adversarial; security analysis; support vector data description; computer security; network intrusion detection; NOVELTY DETECTION; CLASSIFICATION; SUPPORT; ATTACKS; SELF;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Security issues are crucial in a number of machine learning applications, especially in scenarios dealing with human activity rather than natural phenomena (e. g., information ranking, spam detection, malware detection, etc.). In such cases, learning algorithms may have to cope with manipulated data aimed at hampering decision making. Although some previous work addressed the issue of handling malicious data in the context of supervised learning, very little is known about the behavior of anomaly detection methods in such scenarios. In this contribution,(1) we analyze the performance of a particular method-online centroid anomaly detection-in the presence of adversarial noise. Our analysis addresses the following security-related issues: formalization of learning and attack processes, derivation of an optimal attack, and analysis of attack efficiency and limitations. We derive bounds on the effectiveness of a poisoning attack against centroid anomaly detection under different conditions: attacker's full or limited control over the traffic and bounded false positive rate. Our bounds show that whereas a poisoning attack can be effectively staged in the unconstrained case, it can be made arbitrarily difficult (a strict upper bound on the attacker's gain) if external constraints are properly used. Our experimental evaluation, carried out on real traces of HTTP and exploit traffic, confirms the tightness of our theoretical bounds and the practicality of our protection mechanisms.
引用
收藏
页码:3681 / 3724
页数:44
相关论文
共 70 条
[1]  
Angluin D., 1988, MACH LEARN, V2, P434
[2]  
[Anonymous], 2006, P 23 INT C MACHINE, DOI DOI 10.1145/1143844.1143889
[3]  
[Anonymous], 2004, KERNEL METHODS PATTE
[4]  
[Anonymous], P DIMVA
[5]  
[Anonymous], P SIAM INT C DAT MIN
[6]  
[Anonymous], 2008, Exploiting Machine Learning to Subvert Your Spam Filter
[7]  
[Anonymous], C EM ANT
[8]   Learning nested differences in the presence of malicious noise [J].
Auer, P .
THEORETICAL COMPUTER SCIENCE, 1997, 185 (01) :159-175
[9]  
Bailey M, 2007, LECT NOTES COMPUT SC, V4637, P178
[10]  
Barreno M., 2006, P 2006 ACM S INFORM, P16