Facilitating Early-Stage Backdoor Attacks in Federated Learning With Whole Population Distribution Inference

被引:2
作者
Liu, Tian [1 ,2 ]
Hu, Xueyang [3 ]
Shu, Tao [3 ]
机构
[1] Intelligent Network Res Inst, Zhejiang Lab, Hangzhou 311121, Peoples R China
[2] DBAPPSecur Co Ltd, AiLPHA Prod Line Big Data Intelligent Secur, Hangzhou 310051, Peoples R China
[3] Auburn Univ, Dept Comp Sci & Software Engn, Auburn, AL 36849 USA
基金
美国国家科学基金会;
关键词
Internet of Things; Training; Convergence; Statistics; Sociology; Hidden Markov models; Data models; Backdoor attack; federated learning (FL); privacy leakage; weight divergence;
D O I
10.1109/JIOT.2023.3237806
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of the Internet of Things (IoT) combined with the emergence of federated learning (FL) makes it possible for mobile edge computing (MEC) to gain insight from physically separated data without violating privacy or burdening communication. Due to the distributed nature of MEC devices, researchers have uncovered that the FL is vulnerable to backdoor attacks, which aim at injecting a subtask into the FL without corrupting the performance of the main task. The backdoor attack achieves high accuracy on both the main task and the backdoor subtask when injected at FL model convergence. However, the effectiveness of the backdoor is weak when injected in early training stage. In this article, we strengthen the early-injected backdoor attack by using information leakage. We show that FL convergence can be expedited if the client's data set mimics the distribution and gradients of the whole population. Based on this observation, we propose a two-phase backdoor attack, which includes a preliminary phase for the subsequent backdoor attack. Taking advantage of the preliminary phase, the later injected backdoor achieves better effectiveness, as the backdoor effect is less likely to be diluted by normal model updates. Extensive experiments are conducted on the MNIST data set under various data heterogeneity settings to evaluate the effectiveness of the proposed backdoor attack. The results show that the proposed backdoor outperforms existing backdoor attacks in both success rate and longevity, even when defense mechanisms are in place.
引用
收藏
页码:10385 / 10399
页数:15
相关论文
共 52 条
  • [1] Deep Learning with Differential Privacy
    Abadi, Martin
    Chu, Andy
    Goodfellow, Ian
    McMahan, H. Brendan
    Mironov, Ilya
    Talwar, Kunal
    Zhang, Li
    [J]. CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, : 308 - 318
  • [2] Ateniese Giuseppe, 2015, International Journal of Security and Networks, V10, P137
  • [3] Bagdasaryan E., 2018, arXiv
  • [4] Barnett T., 2018, CISCO VISUAL NETWORK, P1
  • [5] Blanchard P, 2017, ADV NEUR IN, V30
  • [6] Bonawitz K., 2016, arXiv, DOI DOI 10.48550/ARXIV.1611.04482
  • [7] Federated learning of predictive models from federated Electronic Health Records
    Brisimi, Theodora S.
    Chen, Ruidi
    Mela, Theofanie
    Olshevsky, Alex
    Paschalidis, Ioannis Ch.
    Shi, Wei
    [J]. INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, 2018, 112 : 59 - 67
  • [8] Sample size selection in optimization methods for machine learning
    Byrd, Richard H.
    Chin, Gillian M.
    Nocedal, Jorge
    Wu, Yuchen
    [J]. MATHEMATICAL PROGRAMMING, 2012, 134 (01) : 127 - 155
  • [9] Chen MQ, 2019, Arxiv, DOI arXiv:1903.10635
  • [10] Security and Privacy-Enhanced Federated Learning for Anomaly Detection in IoT Infrastructures
    Cui, Lei
    Qu, Youyang
    Xie, Gang
    Zeng, Deze
    Li, Ruidong
    Shen, Shigen
    Yu, Shui
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (05) : 3492 - 3500