Subpopulation Data Poisoning Attacks

被引:32
作者
Jagielski, Matthew [1 ]
Severi, Giorgio [1 ]
Harger, Niklas Pousette [1 ]
Oprea, Mina [1 ]
机构
[1] Northeastern Univ, Boston, MA 02115 USA
来源
CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY | 2021年
关键词
Adversarial Machine Learning; Poisoning Attacks; Fairness;
D O I
10.1145/3460120.3485368
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning systems are deployed in critical settings, but they might fail in unexpected ways, impacting the accuracy of their predictions. Poisoning attacks against machine learning induce adversarial modification of data used by a machine learning algorithm to selectively change its output when it is deployed. In this work, we introduce a novel data poisoning attack called a subpopulation attack, which is particularly relevant when datasets are large and diverse. We design a modular framework for subpopulation attacks, instantiate it with different building blocks, and show that the attacks are effective for a variety of datasets and machine learning models. We further optimize the attacks in continuous domains using influence functions and gradient optimization methods. Compared to existing backdoor poisoning attacks, subpopulation attacks have the advantage of inducing misclassification in naturally distributed data points at inference time, making the attacks extremely stealthy. We also show that our attack strategy can be used to improve upon existing targeted attacks. We prove that, under some assumptions, subpopulation attacks are impossible to defend against, and empirically demonstrate the limitations of existing defenses against our attacks, highlighting the difficulty of protecting machine learning against this threat.
引用
收藏
页码:3104 / 3122
页数:19
相关论文
共 81 条
  • [1] [Anonymous], 2020, P IEEE S SECUR PRIV, DOI DOI 10.1109/SP40000.2020.00115
  • [2] [Anonymous], 2015, INT C MACH LEARN
  • [3] [Anonymous], P 29 AAAI C ART INT
  • [4] [Anonymous], 2018, P 27 US SEC S
  • [5] Bagdasaryan E, 2020, PR MACH LEARN RES, V108, P2938
  • [6] Basu Samyadeep, 2021, INT C LEARN REPR
  • [7] Beaufays F., 2018, ARXIV181103604
  • [8] Biggio B., 2014, Proceedings of the 2014 Workshop on Artificial Intelligent and Security Workshop - AISec'14, P27, DOI DOI 10.1145/2666652.2666666
  • [9] Biggio B., 2012, P 29 INT C MACH LEAR
  • [10] Biggio B., 2013, LNCS, P387, DOI [DOI 10.1007/978-3-642-40994-325, DOI 10.1007/978-3-642-40994-3_25]