BadCleaner: Defending Backdoor Attacks in Federated Learning via Attention-Based Multi-Teacher Distillation

被引:7
作者
Zhang, Jiale [1 ]
Zhu, Chengcheng [1 ]
Ge, Chunpeng [2 ]
Ma, Chuan [3 ]
Zhao, Yanchao [4 ]
Sun, Xiaobing [1 ]
Chen, Bing [4 ]
机构
[1] Yangzhou Univ, Sch Informat Engn, Yangzhou 225127, Peoples R China
[2] Shandong Univ, Sch Software, Jinan 250000, Peoples R China
[3] Zhejiang Lab, Hangzhou 311100, Peoples R China
[4] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 211106, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Data models; Training; Germanium; Federated learning; Degradation; Watermarking; Training data; backdoor attacks; multi-teacher distillation; attention transfer;
D O I
10.1109/TDSC.2024.3354049
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
As a privacy-preserving distributed learning paradigm, federated learning (FL) has been proven to be vulnerable to various attacks, among which backdoor attack is one of the toughest. In this attack, malicious users attempt to embed backdoor triggers into local models, resulting in the crafted inputs being misclassified as the targeted labels. To address such attack, several defense mechanisms are proposed, but may lose the effectiveness due to the following drawbacks. First, current methods heavily rely on massive labeled clean data, which is an impractical setting in FL. Moreover, an in-avoidable performance degradation usually occurs in the defensive procedure. To alleviate such concerns, we propose BadCleaner, a lossless and efficient backdoor defense scheme via attention-based federated multi-teacher distillation. First, BadCleaner can effectively tune the backdoored joint model without performance degradation, by distilling the in-depth knowledge from multiple teachers with only a small part of unlabeled clean data. Second, to fully eliminate the hidden backdoor patterns, we present an attention transfer method to alleviate the attention of models to the trigger regions. The extensive evaluation demonstrates that BadCleaner can reduce the success rates of state-of-the-art backdoor attacks without compromising the model performance.
引用
收藏
页码:4559 / 4573
页数:15
相关论文
共 56 条
[31]  
Rong DZ, 2022, PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, P2204
[32]   Privacy-Preserving Deep Learning [J].
Shokri, Reza ;
Shmatikov, Vitaly .
CCS'15: PROCEEDINGS OF THE 22ND ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2015, :1310-1321
[33]  
Nguyen TD, 2022, PROCEEDINGS OF THE 31ST USENIX SECURITY SYMPOSIUM, P1415
[34]  
Tran B, 2018, ADV NEUR IN, V31
[35]   Neural Cleanse: Identifying and Mitigating Backdoor Attacks in Neural Networks [J].
Wang, Bolun ;
Yao, Yuanshun ;
Shan, Shawn ;
Li, Huiying ;
Viswanath, Bimal ;
Zheng, Haitao ;
Zhao, Ben Y. .
2019 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2019), 2019, :707-723
[36]  
Wang H, 2020, Adv. Neural Inf. Process. Syst.
[37]   RAB: Provable Robustness Against Backdoor Attacks [J].
Weber, Maurice ;
Xu, Xiaojun ;
Karlas, Bojan ;
Zhang, Ce ;
Li, Bo .
2023 IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP, 2023, :1311-1328
[38]   Communication-efficient federated learning via knowledge distillation [J].
Wu, Chuhan ;
Wu, Fangzhao ;
Lyu, Lingjuan ;
Huang, Yongfeng ;
Xie, Xing .
NATURE COMMUNICATIONS, 2022, 13 (01)
[39]  
Xia J, 2022, PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, P1481
[40]  
Xia PF, 2022, PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, P3992