BadCleaner: Defending Backdoor Attacks in Federated Learning via Attention-Based Multi-Teacher Distillation

被引:7
作者
Zhang, Jiale [1 ]
Zhu, Chengcheng [1 ]
Ge, Chunpeng [2 ]
Ma, Chuan [3 ]
Zhao, Yanchao [4 ]
Sun, Xiaobing [1 ]
Chen, Bing [4 ]
机构
[1] Yangzhou Univ, Sch Informat Engn, Yangzhou 225127, Peoples R China
[2] Shandong Univ, Sch Software, Jinan 250000, Peoples R China
[3] Zhejiang Lab, Hangzhou 311100, Peoples R China
[4] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 211106, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Data models; Training; Germanium; Federated learning; Degradation; Watermarking; Training data; backdoor attacks; multi-teacher distillation; attention transfer;
D O I
10.1109/TDSC.2024.3354049
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
As a privacy-preserving distributed learning paradigm, federated learning (FL) has been proven to be vulnerable to various attacks, among which backdoor attack is one of the toughest. In this attack, malicious users attempt to embed backdoor triggers into local models, resulting in the crafted inputs being misclassified as the targeted labels. To address such attack, several defense mechanisms are proposed, but may lose the effectiveness due to the following drawbacks. First, current methods heavily rely on massive labeled clean data, which is an impractical setting in FL. Moreover, an in-avoidable performance degradation usually occurs in the defensive procedure. To alleviate such concerns, we propose BadCleaner, a lossless and efficient backdoor defense scheme via attention-based federated multi-teacher distillation. First, BadCleaner can effectively tune the backdoored joint model without performance degradation, by distilling the in-depth knowledge from multiple teachers with only a small part of unlabeled clean data. Second, to fully eliminate the hidden backdoor patterns, we present an attention transfer method to alleviate the attention of models to the trigger regions. The extensive evaluation demonstrates that BadCleaner can reduce the success rates of state-of-the-art backdoor attacks without compromising the model performance.
引用
收藏
页码:4559 / 4573
页数:15
相关论文
共 56 条
[1]   Variational Information Distillation for Knowledge Transfer [J].
Ahn, Sungsoo ;
Hu, Shell Xu ;
Damianou, Andreas ;
Lawrence, Neil D. ;
Dai, Zhenwen .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9155-9163
[2]   BaFFLe: Backdoor Detection via Feedback -based Federated Learning [J].
Andreina, Sebastien ;
Marson, Giorgia Azzurra ;
Moellering, Helen ;
Karame, Ghassan .
2021 IEEE 41ST INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2021), 2021, :852-863
[3]   CONTRA: Defending Against Poisoning Attacks in Federated Learning [J].
Awan, Sana ;
Luo, Bo ;
Li, Fengjun .
COMPUTER SECURITY - ESORICS 2021, PT I, 2021, 12972 :455-475
[4]  
Bagdasaryan E, 2020, PR MACH LEARN RES, V108, P2938
[5]  
Bhagoji AN, 2019, PR MACH LEARN RES, V97
[6]   Practical Secure Aggregation for Privacy-Preserving Machine Learning [J].
Bonawitz, Keith ;
Ivanov, Vladimir ;
Kreuter, Ben ;
Marcedone, Antonio ;
McMahan, H. Brendan ;
Patel, Sarvar ;
Ramage, Daniel ;
Segal, Aaron ;
Seth, Karn .
CCS'17: PROCEEDINGS OF THE 2017 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2017, :1175-1191
[7]   Understanding Distributed Poisoning Attack in Federated Learning [J].
Cao, Di ;
Chang, Shan ;
Lin, Zhijian ;
Liu, Guohua ;
Sunt, Donghong .
2019 IEEE 25TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2019, :233-239
[8]  
Fang MH, 2020, PROCEEDINGS OF THE 29TH USENIX SECURITY SYMPOSIUM, P1623
[9]   SAFELearn: Secure Aggregation for private FEderated Learning [J].
Fereidooni, Hossein ;
Marchal, Samuel ;
Miettinen, Markus ;
Mirhoseini, Azalia ;
Moellering, Helen ;
Thien Duc Nguyen ;
Rieger, Phillip ;
Sadeghi, Ahmad-Reza ;
Schneider, Thomas ;
Yalame, Hossein ;
Zeitouni, Shaza .
2021 IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2021), 2021, :56-62
[10]  
Fung Clement, 2020, 23 INT S RES ATTACKS, P301