PDLM: Privacy-Preserving Deep Learning Model on Cloud with Multiple Keys

被引:37
作者
Ma, Xindi [1 ]
Ma, Jianfeng [1 ]
Li, Hui [1 ]
Jiang, Qi [1 ]
Gao, Sheng [2 ]
机构
[1] Xidian Univ, Sch Cyber Engn, Xian 710071, Shaanxi, Peoples R China
[2] Cent Univ Finance & Econ, Sch Informat, Beijing 102202, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Cryptography; Machine learning; Training; Privacy; Data models; Data privacy; Computational modeling; Privacy preservation; deep learning; cryptography; multiple keys; TRAJECTORY PRIVACY; FRAMEWORK; MECHANISM; CROWD;
D O I
10.1109/TSC.2018.2868750
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has aroused a lot of attention and has been used successfully in many domains, such as accurate image recognition and medical diagnosis. Generally, the training of models requires large, representative datasets, which may be collected from a large number of users and contain sensitive information (e.g., users' photos and medical information). The collected data would be stored and computed by service providers (SPs) or delegated to an untrusted cloud. The users can neither control how it will be used, nor realize what will be learned from it, which make the privacy issues prominent and severe. To solve the privacy issues, one of the most popular approaches is to encrypt users' data with their public keys. However, this technique inevitably leads to another challenge that how to train the model based on multi-key encrypted data. In this paper, we propose a novel privacy-preserving deep learning model, namely PDLM, to apply deep learning over the encrypted data under multiple keys. In PDLM, lots of users contribute their encrypted data to SP to learn a specific model. We adopt an effective privacy-preserving calculation toolkit to achieve the training process based on stochastic gradient descent (SGD) in a privacy-preserving manner. We also prove that our PDLM can achieve users' privacy preservation and analyze the efficiency of PDLM in theory. Finally, we conduct an experiment to evaluate PDLM over two real-world datasets and empirical results demonstrate that our PDLM can effectively and efficiently train the model in a privacy-preserving way.
引用
收藏
页码:1251 / 1263
页数:13
相关论文
共 50 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]  
Agrawal R, 2000, SIGMOD REC, V29, P439, DOI 10.1145/335191.335438
[3]  
[Anonymous], 2012, P 6 ACM C RECOMMENDE
[4]  
Avriel M., 2003, NONLINEAR PROGRAMMIN
[5]   Machine Learning Classification over Encrypted Data [J].
Bost, Raphael ;
Popa, Raluca Ada ;
Tu, Stephen ;
Goldwasser, Shafi .
22ND ANNUAL NETWORK AND DISTRIBUTED SYSTEM SECURITY SYMPOSIUM (NDSS 2015), 2015,
[6]  
Bresson E, 2003, LECT NOTES COMPUT SC, V2894, P37
[7]  
Collobert Ronan, 2011, BIGLEARN NIPS WORKSH
[8]   A Forensically Sound Adversary Model for Mobile Devices [J].
Do, Quang ;
Martini, Ben ;
Choo, Kim-Kwang Raymond .
PLOS ONE, 2015, 10 (09)
[9]  
Du WL, 2004, SIAM PROC S, P222
[10]  
Dwork C, 2006, LECT NOTES COMPUT SC, V4052, P1