Resisting membership inference attacks through knowledge distillation

被引:19
作者
Zheng, Junxiang [1 ]
Cao, Yongzhi [1 ]
Wang, Hanpin [1 ,2 ]
机构
[1] Peking Univ, Dept Comp Sci & Technol, Key Lab High Confidence Software Technol MOE, Beijing 100871, Peoples R China
[2] Guangzhou Univ, Sch Comp Sci & Cyber Engn, Guangzhou 510006, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Privacy; Membership inference attacks; Knowledge distillation; Membership privacy; Deep learning;
D O I
10.1016/j.neucom.2021.04.082
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, membership inference attacks (MIAs) against machine learning models have been proposed. Using MIAs, adversaries can inference whether a data record is in the training set of the target model. Defense methods which use differential privacy mechanisms or adversarial training cannot handle the trade-off between privacy and utility well. Other methods based on knowledge transfer to improve model utility need public unlabeled data in the same distribution as private data, and this requirement may not be satisfied in some scenarios. To handle the trade-off between privacy and utility better, we propose two algorithms of deep learning, i.e., complementary knowledge distillation (CKD) and pseudo complementary knowledge distillation (PCKD). In CKD, the transfer data of knowledge distillation all come from the private training set, but their soft targets are generated from the teacher model which is trained using their complementary set. With similar idea, we propose PCKD which reduces the training set of each teacher model and uses model averaging to generate soft targets of transfer data. Because smaller training set leads to less utility, PCKD utilizes pre-training to improve the utility of teacher models. Experimental results on widely used datasets show that CKD and PCKD can both averagely decrease attack accuracy by nearly 25% with negligible utility loss. The training time of PCKD is nearly 40% lower than that of CKD. Compared with existing defense methods such as DMP, adversarial regularization, dropout, and DPSGD, CKD and PCKD have great advantages on handling the trade-off between privacy and utility. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:114 / 126
页数:13
相关论文
共 34 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]   A review of privacy-preserving techniques for deep learning [J].
Boulemtafes, Amine ;
Derhab, Abdelouahid ;
Challal, Yacine .
NEUROCOMPUTING, 2020, 384 :21-45
[3]  
Chaudhuri K, 2011, J MACH LEARN RES, V12, P1069
[4]  
Dwork C, 2006, LECT NOTES COMPUT SC, V4052, P1
[5]   The Algorithmic Foundations of Differential Privacy [J].
Dwork, Cynthia ;
Roth, Aaron .
FOUNDATIONS AND TRENDS IN THEORETICAL COMPUTER SCIENCE, 2013, 9 (3-4) :211-406
[6]  
Furlanello T., 2018, PR MACH LEARN RES, P1607
[7]  
Hayes Jamie, 2019, Proceedings on Privacy Enhancing Technologies, V2019, P133, DOI 10.2478/popets-2019-0008
[8]  
Hinton G., 2015, ARXIV
[9]   Densely Connected Convolutional Networks [J].
Huang, Gao ;
Liu, Zhuang ;
van der Maaten, Laurens ;
Weinberger, Kilian Q. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2261-2269
[10]  
Ioffe S., 2015, PMLR, P448, DOI DOI 10.48550/ARXIV.1502.03167