Towards Deep Neural Network Training on Encrypted Data

被引:77
作者
Nandakumar, Karthik [1 ]
Ratha, Nalini [2 ]
Pankanti, Sharath [2 ]
Halevi, Shai [3 ]
机构
[1] IBM Res, Singapore, Singapore
[2] IBM Res, Yorktown Hts, NY 10598 USA
[3] Algorand Fdn, Boston, MA USA
来源
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019) | 2019年
关键词
FULLY HOMOMORPHIC ENCRYPTION;
D O I
10.1109/CVPRW.2019.00011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While deep learning is a valuable tool for solving many tough problems in computer vision., the success of deep learning models is typically determined by: (1) availability of sufficient training data, (ii) access to extensive computational resources, and (iii) expertise in selecting the right model and hyperparameters for the selected task. Often, the availability of data is the hard part due to compliance, legal, and privacy constraints. Cryptographic techniques such as fully homomorphic encryption, (THE) offer a potential solution by enabling processing on encrypted data. While prior work has been done on using FHE for inferencing, training a deep neural network in the encrypted domain is an extremely challenging task due to the computational complexity of the operations involved. In this paper, we evaluate the feasibility of training neural networks on encrypted data in a completely non-interactive way. Our proposed system uses the open-source FHE toolkit HElib to implement a Stochastic Gradient Descent (SGD)-based training of a neural network. We show that encrypted training can be made more computationally efficient by (i) simplifyinj_s, the network with minimal degradation of accuracy, choosing appropriate data representation and resolution, and (iii) packing the data elements within the ciphertext in a smart way so as to minimize the number of operations and facilitate parallelization of FHE computations. Based on the above optimizations, we demonstrate that it is possible to achieve more than 50x speed up while training a fully-connected neural network on the MNIST dataset while achieving reasonable accuracy (96%). Though the cost of training a complex deep learning model from scratch on encrypted data is still very high, this work establishes a solid baseline and paves the way for relatively simpler tasks such as fine-tuning of deep learning models based on encrypted data to be implemented in the near future.
引用
收藏
页码:40 / 48
页数:9
相关论文
共 30 条
  • [1] Agrawal R, 2000, SIGMOD REC, V29, P439, DOI 10.1145/335191.335438
  • [2] [Anonymous], HElib: An Implementation of homomorphic encryption
  • [3] [Anonymous], 2018, 2018254 CRYPT EPRINT
  • [4] Barni M., 2006, P 8 WORKSH MULT SEC, P146
  • [5] Boddeti V. N., 2018, ARXIV180500577
  • [6] Boneh Dan, 2013, PROC INT C APPL CRY, P102, DOI 10.1007/978-3-642-38980-1_7
  • [7] Brakerski Zvika, 2014, ACM Transactions on Computation Theory, V6, DOI 10.1145/2633600
  • [8] Fully Homomorphic Encryption without Modulus Switching from Classical GapSVP
    Brakerski, Zvika
    [J]. ADVANCES IN CRYPTOLOGY - CRYPTO 2012, 2012, 7417 : 868 - 886
  • [9] Chialva D., 2018, 20181032 CRYPT EPRIN
  • [10] Costache Anamaria, 2017, Selected Areas in Cryptography - SAC 2016. 23rd International Conference. Revised Selected Papers: LNCS 10532, P401, DOI 10.1007/978-3-319-69453-5_22