Local Differential Privacy for Deep Learning

被引:155
作者
Arachchige, Pathum Chamikara Mahawaga [1 ,2 ]
Bertok, Peter [1 ]
Khalil, Ibrahim [1 ]
Liu, Dongxi [2 ]
Camtepe, Seyit [2 ]
Atiquzzaman, Mohammed [3 ]
机构
[1] RMIT Univ, Sch Sci, Dept Comp Sci & Software Engn, Melbourne, Vic 3000, Australia
[2] CSIRO Data61, Melbourne, Vic 3008, Australia
[3] Univ Oklahoma, Sch Comp Sci, Norman, OK 73019 USA
来源
IEEE INTERNET OF THINGS JOURNAL | 2020年 / 7卷 / 07期
关键词
Data privacy; deep learning (DL); differential privacy (DP); local DP (LDP); CLOUD;
D O I
10.1109/JIOT.2019.2952146
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The Internet of Things (IoT) is transforming major industries, including but not limited to healthcare, agriculture, finance, energy, and transportation. IoT platforms are continually improving with innovations, such as the amalgamation of software-defined networks (SDNs) and network function virtualization (NFV) in the edge-cloud interplay. Deep learning (DL) is becoming popular due to its remarkable accuracy when trained with a massive amount of data such as generated by IoT. However, DL algorithms tend to leak privacy when trained on highly sensitive crowd-sourced data such as medical data. The existing privacy-preserving DL algorithms rely on the traditional server-centric approaches requiring high processing powers. We propose a new local differentially private (LDP) algorithm named LATENT that redesigns the training process. LATENT enables a data owner to add a randomization layer before data leave the data owners' devices and reach a potentially untrusted machine learning service. This feature is achieved by splitting the architecture of a convolutional neural network (CNN) into three layers: 1) convolutional module (CNM); 2) randomization module; and 3) fully connected module. Hence, the randomization module can operate as an NFV privacy preservation service in an SDN-controlled NFV, making LATENT more practical for IoT-driven cloud-based environments compared to existing approaches. The randomization module employs a newly proposed LDP protocol named utility enhancing randomization, which allows LATENT to maintain high utility compared to existing LDP protocols. Our experimental evaluation of LATENT on convolutional deep neural networks demonstrates excellent accuracy (e.g., 91%-96%) with high model quality even under low privacy budgets (e.g., epsilon=0.5).
引用
收藏
页码:5827 / 5842
页数:16
相关论文
共 51 条
[1]  
Abadi M., 2016, TENSORFLOW LARGE SCA
[2]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[3]  
[Anonymous], 2015, Randomized response and related methods: Surveying sensitive data
[4]  
[Anonymous], 2017, ARXIV170302952
[5]   Local, Private, Efficient Protocols for Succinct Histograms [J].
Bassily, Raef ;
Smith, Adam .
STOC'15: PROCEEDINGS OF THE 2015 ACM SYMPOSIUM ON THEORY OF COMPUTING, 2015, :127-135
[6]   Concentrated Differential Privacy: Simplifications, Extensions, and Lower Bounds [J].
Bun, Mark ;
Steinke, Thomas .
THEORY OF CRYPTOGRAPHY, TCC 2016-B, PT I, 2016, 9985 :635-658
[7]   An efficient and scalable privacy preserving algorithm for big data and data streams [J].
Chamikara, M. A. P. ;
Bertok, P. ;
Liu, D. ;
Camtepe, S. ;
Khalil, I .
COMPUTERS & SECURITY, 2019, 87
[8]   Efficient privacy preservation of big data for accurate data mining [J].
Chamikara, M. A. P. ;
Bertok, P. ;
Liu, D. ;
Camtepe, S. ;
Khalil, I .
INFORMATION SCIENCES, 2020, 527 :420-443
[9]   Efficient data perturbation for privacy preserving and accurate data stream mining [J].
Chamikara, M. A. P. ;
Bertok, P. ;
Liu, D. ;
Camtepe, S. ;
Khalil, I .
PERVASIVE AND MOBILE COMPUTING, 2018, 48 :1-19
[10]  
Chan T.-H Hubert, 2012, Privacy Enhancing Technologies. Proceedings 12th International Symposium, PETS 2012, P140, DOI 10.1007/978-3-642-31680-7_8