Evil vs evil: using adversarial examples to against backdoor attack in federated learning

被引:4
作者
Liu, Tao [1 ]
Li, Mingjun [1 ]
Zheng, Haibin [1 ,2 ]
Ming, Zhaoyan [3 ]
Chen, Jinyin [1 ,2 ]
机构
[1] Zhejiang Univ Technol, Coll Informat Engn, Hangzhou 310023, Peoples R China
[2] Zhejiang Univ Technol, Inst Cyberspace Secur, Hangzhou 310023, Peoples R China
[3] Zhejiang Univ City Coll, Coll Comp & Comp Sci, Hangzhou 310015, Peoples R China
基金
中国国家自然科学基金;
关键词
Federated learning; Backdoor attacks; Defense; Adversarial attacks; Clustering algorithm; SECURITY; PRIVACY;
D O I
10.1007/s00530-022-00965-z
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As a distributed learning paradigm, federated learning (FL) has shown great success in aggregating information from different clients to train a shared global model. Unfortunately, by uploading carefully crafted updated models, a malicious client can embed a backdoor into the global model during FL's training. Numerous secure aggregation strategies and robust training protocols have been proposed to defend FL against backdoor attacks. However, they are still challenged, either being bypassed by adaptive attacks or sacrificing the main task performance of FL. By conducting empirical studies of backdoor attacks in FL, we gain an interesting insight that adversarial perturbations can activate backdoors in backdoor models. Consequently, behavior differences of models fed by adversarial examples are compared for backdoor update detection. We propose a novel FL backdoor defense method using adversarial examples, denoted as Evil vs Evil (EVE). Specifically, a small data set of clean examples for FL's main task training is collected in the sever for adversarial examples generation. By observing the behavior of updated models under the adversarial examples, EVE uses a clustering algorithm to select benign models and to exclude the other models, without any loss of the main task performance of FL itself. Extensive evaluations across four data sets and the corresponding DNNs demonstrate the state-of-the-art (SOTA) defense performance of EVE compared with five baselines. In particular, EVE under 40% of malicious clients can reduce the attack success rate from 99% to 1%. In addition, we verify that EVE is still robust under the adaptive attacks. EVE is open sourced to facilitate future research.
引用
收藏
页码:553 / 568
页数:16
相关论文
共 52 条
[1]  
Aledhari M, 2020, IEEE ACCESS, V8, P140699, DOI [10.1109/ACCESS.2020.3013541, 10.1109/access.2020.3013541]
[2]   BaFFLe: Backdoor Detection via Feedback -based Federated Learning [J].
Andreina, Sebastien ;
Marson, Giorgia Azzurra ;
Moellering, Helen ;
Karame, Ghassan .
2021 IEEE 41ST INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2021), 2021, :852-863
[3]  
[Anonymous], 2018, CoRR
[4]  
Bagdasaryan E, 2020, PR MACH LEARN RES, V108, P2938
[5]  
Bhagoji AN, 2019, PR MACH LEARN RES, V97
[6]  
Blanchard P, 2017, ADV NEUR IN, V30
[7]  
Brendel Wieland, 2018, P INT C LEARN REPR
[8]   FLTrust: Byzantine-robust Federated Learning via Trust Bootstrapping [J].
Cao, Xiaoyu ;
Fang, Minghong ;
Liu, Jia ;
Gong, Neil Zhenqiang .
28TH ANNUAL NETWORK AND DISTRIBUTED SYSTEM SECURITY SYMPOSIUM (NDSS 2021), 2021,
[9]  
Chang T., 2018, CORR 2018 ARXIV18100
[10]  
Chen B., 2019, P 10 ACM MULTIMEDIA, V2301