Evil vs evil: using adversarial examples to against backdoor attack in federated learning

被引:4
作者
Liu, Tao [1 ]
Li, Mingjun [1 ]
Zheng, Haibin [1 ,2 ]
Ming, Zhaoyan [3 ]
Chen, Jinyin [1 ,2 ]
机构
[1] Zhejiang Univ Technol, Coll Informat Engn, Hangzhou 310023, Peoples R China
[2] Zhejiang Univ Technol, Inst Cyberspace Secur, Hangzhou 310023, Peoples R China
[3] Zhejiang Univ City Coll, Coll Comp & Comp Sci, Hangzhou 310015, Peoples R China
基金
中国国家自然科学基金;
关键词
Federated learning; Backdoor attacks; Defense; Adversarial attacks; Clustering algorithm; SECURITY; PRIVACY;
D O I
10.1007/s00530-022-00965-z
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As a distributed learning paradigm, federated learning (FL) has shown great success in aggregating information from different clients to train a shared global model. Unfortunately, by uploading carefully crafted updated models, a malicious client can embed a backdoor into the global model during FL's training. Numerous secure aggregation strategies and robust training protocols have been proposed to defend FL against backdoor attacks. However, they are still challenged, either being bypassed by adaptive attacks or sacrificing the main task performance of FL. By conducting empirical studies of backdoor attacks in FL, we gain an interesting insight that adversarial perturbations can activate backdoors in backdoor models. Consequently, behavior differences of models fed by adversarial examples are compared for backdoor update detection. We propose a novel FL backdoor defense method using adversarial examples, denoted as Evil vs Evil (EVE). Specifically, a small data set of clean examples for FL's main task training is collected in the sever for adversarial examples generation. By observing the behavior of updated models under the adversarial examples, EVE uses a clustering algorithm to select benign models and to exclude the other models, without any loss of the main task performance of FL itself. Extensive evaluations across four data sets and the corresponding DNNs demonstrate the state-of-the-art (SOTA) defense performance of EVE compared with five baselines. In particular, EVE under 40% of malicious clients can reduce the attack success rate from 99% to 1%. In addition, we verify that EVE is still robust under the adaptive attacks. EVE is open sourced to facilitate future research.
引用
收藏
页码:553 / 568
页数:16
相关论文
共 52 条
[11]  
Choudhury Olivia, 2019, AMIA Annu Symp Proc, V2019, P313
[12]   Boosting Adversarial Attacks with Momentum [J].
Dong, Yinpeng ;
Liao, Fangzhou ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun ;
Hu, Xiaolin ;
Li, Jianguo .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9185-9193
[13]  
El Mhamdi E. M., 2018, PR MACH LEARN RES, P3518
[14]  
Fang MH, 2020, PROCEEDINGS OF THE 29TH USENIX SECURITY SYMPOSIUM, P1623
[15]   Clustering by passing messages between data points [J].
Frey, Brendan J. ;
Dueck, Delbert .
SCIENCE, 2007, 315 (5814) :972-976
[16]  
Fu S., 2019, CORR 2019 ARXIV19121
[17]  
Fung C., 2020, P 23 INT S RES ATT I, P301
[18]  
Goodfellow I., 2015, CoRR
[19]   BadNets: Evaluating Backdooring Attacks on Deep Neural Networks [J].
Gu, Tianyu ;
Liu, Kang ;
Dolan-Gavitt, Brendan ;
Garg, Siddharth .
IEEE ACCESS, 2019, 7 :47230-47244
[20]  
HE KM, 2016, PROC CVPR IEEE, P770, DOI [DOI 10.1109/CVPR.2016.90, 10.1109/CVPR.2016.90]