De-Pois: An Attack-Agnostic Defense against Data Poisoning Attacks

被引:60
作者
Chen, Jian [1 ]
Zhang, Xuxin [1 ]
Zhang, Rui [2 ]
Wang, Chen [1 ]
Liu, Ling [3 ]
机构
[1] Huazhong Univ Sci & Technol, Internet Technol & Engn Res & Dev Ctr ITEC, Sch Elect Informat & Commun, Wuhan 430074, Peoples R China
[2] Wuhan Univ Technol, Sch Comp Sci & Technol, Hubei Key Lab Transportat Internet Things, Wuhan 430070, Peoples R China
[3] Georgia Inst Technol, Coll Comp, Atlanta, GA 30332 USA
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Data models; Training; Testing; Predictive models; Computational modeling; Training data; Task analysis; Machine learning; data poisoning attack; attack-agnostic defense; generative adversarial network;
D O I
10.1109/TIFS.2021.3080522
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Machine learning techniques have been widely applied to various applications. However, they are potentially vulnerable to data poisoning attacks, where sophisticated attackers can disrupt the learning procedure by injecting a fraction of malicious samples into the training dataset. Existing defense techniques against poisoning attacks are largely attack-specific: they are designed for one specific type of attacks but do not work for other types, mainly due to the distinct principles they follow. Yet few general defense strategies have been developed. In this paper, we propose De-Pois, an attack-agnostic defense against poisoning attacks. The key idea of De-Pois is to train a mimic model the purpose of which is to imitate the behavior of the target model trained by clean samples. We take advantage of Generative Adversarial Networks (GANs) to facilitate informative training data augmentation as well as the mimic model construction. By comparing the prediction differences between the mimic model and the target model, De-Pois is thus able to distinguish the poisoned samples from clean ones, without explicit knowledge of any ML algorithms or types of poisoning attacks. We implement four types of poisoning attacks and evaluate De-Pois with five typical defense methods on different realistic datasets. The results demonstrate that De-Pois is effective and efficient for detecting poisoned data against all the four types of poisoning attacks, with both the accuracy and F1-score over 0.9 on average.
引用
收藏
页码:3412 / 3425
页数:14
相关论文
共 43 条
[11]  
Chen C., 2019, IEEE Trans. Inf. Forensics Security, early access, DOI [10.1109/TIFS.2019.2945198, DOI 10.1109/TIFS.2019.2945198]
[12]   Efficient and Private Scoring of Decision Trees, Support Vector Machines and Logistic Regression Models Based on Pre-Computation [J].
De Cock, Martine ;
Dowsley, Rafael ;
Horst, Caleb ;
Katti, Raj ;
Nascimento, Anderson C. A. ;
Poon, Wing-Sea ;
Truex, Stacey .
IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2019, 16 (02) :217-230
[13]  
Diakonikolas I, 2019, PR MACH LEARN RES, V97
[14]   THE FRECHET DISTANCE BETWEEN MULTIVARIATE NORMAL-DISTRIBUTIONS [J].
DOWSON, DC ;
LANDAU, BV .
JOURNAL OF MULTIVARIATE ANALYSIS, 1982, 12 (03) :450-455
[15]   Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures [J].
Fredrikson, Matt ;
Jha, Somesh ;
Ristenpart, Thomas .
CCS'15: PROCEEDINGS OF THE 22ND ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2015, :1322-1333
[16]   STRIP: A Defence Against Trojan Attacks on Deep Neural Networks [J].
Gao, Yansong ;
Xu, Change ;
Wang, Derui ;
Chen, Shiping ;
Ranasinghe, Damith C. ;
Nepal, Surya .
35TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE (ACSA), 2019, :113-125
[17]  
Goodfellow I., 2020, ADV NEUR IN, V63, P139, DOI [DOI 10.1145/3422622, 10.1145/3422622]
[18]  
Gu Tianyu, 2017, P MACH LEARN COMP SE
[19]  
Gulrajani I., 2017, ADV NEURAL INFPROCES, P5767
[20]   Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning [J].
Hitaj, Briland ;
Ateniese, Giuseppe ;
Perez-Cruz, Fernando .
CCS'17: PROCEEDINGS OF THE 2017 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2017, :603-618