Membership Inference Attacks Against Machine Learning Models

被引:2066
作者
Shokri, Reza [1 ]
Stronati, Marco [1 ,2 ]
Song, Congzheng [1 ]
Shmatikov, Vitaly [1 ]
机构
[1] Cornell Tech, New York, NY 10044 USA
[2] INRIA, Rocquencourt, France
来源
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP) | 2017年
基金
美国国家科学基金会;
关键词
PRIVACY;
D O I
10.1109/SP.2017.41
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We quantitatively investigate how machine learning models leak information about the individual data records on which they were trained. We focus on the basic membership inference attack: given a data record and black-box access to a model, determine if the record was in the model's training dataset. To perform membership inference against a target model, we make adversarial use of machine learning and train our own inference model to recognize differences in the target model's predictions on the inputs that it trained on versus the inputs that it did not train on. We empirically evaluate our inference techniques on classification models trained by commercial "machine learning as a service" providers such as Google and Amazon. Using realistic datasets and classification tasks, including a hospital discharge dataset whose membership is sensitive from the privacy perspective, we show that these models can be vulnerable to membership inference attacks. We then investigate the factors that influence this leakage and evaluate mitigation strategies.
引用
收藏
页码:3 / 18
页数:16
相关论文
共 38 条
  • [1] [Anonymous], 2015, ARXIV
  • [2] [Anonymous], 2015, P S FDN COMP SCI
  • [3] [Anonymous], 2006, TCC
  • [4] [Anonymous], 2009, Learning multiple layers of features from tiny images
  • [5] [Anonymous], CCS
  • [6] [Anonymous], SDM
  • [7] [Anonymous], 2016, Statistical inference considered harmful
  • [8] [Anonymous], NIPS
  • [9] [Anonymous], 2016, USENIX SECURITY
  • [10] [Anonymous], NIPS