Scalable and Efficient Training of Large Convolutional Neural Networks with Differential Privacy

被引:0
作者
Bu, Zhiqi [1 ]
Mao, Jialin [1 ]
Xu, Shiyun [1 ]
机构
[1] Univ Penn, Dept Appl Math & Computat Sci, Philadelphia, PA 19104 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large convolutional neural networks (CNN) can be difficult to train in the differentially private (DP) regime, since the optimization algorithms require a computationally expensive operation, known as the per-sample gradient clipping. We propose an efficient and scalable implementation of this clipping on convolutional layers, termed as the mixed ghost clipping, that significantly eases the private training in terms of both time and space complexities, without affecting the accuracy. The improvement in efficiency is rigorously studied through the first complexity analysis for the mixed ghost clipping and existing DP training algorithms. Extensive experiments on vision classification tasks, with large ResNet, VGG, and Vision Transformers (ViT), demonstrate that DP training with mixed ghost clipping adds 1 similar to 10% memory overhead and < 2x slowdown to the standard non-private training. Specifically, when training VGG19 on CIFAR10, the mixed ghost clipping is 3x faster than state-of-the-art Opacus library with 18x larger maximum batch size. To emphasize the significance of efficient DP training on convolutional layers, we achieve 96.7% accuracy on CIFAR10 and 83.0% on CIFAR100 at epsilon = 1 using BEiT, while the previous best results are 94.8% and 67.4%, respectively. We opensource a privacy engine (https://github.com/woodyx218/private_vision) that implements DP training of CNN (including convolutional ViT) with a few lines of code.
引用
收藏
页数:14
相关论文
共 53 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]  
Ali A, 2021, ROU ST S ASIAN POLIT, P31
[3]  
Bao H., 2021, INT C LEARN REPR
[4]   Towards the cold atom analog false vacuum [J].
Braden, Jonathan ;
Johnson, Matthew C. ;
Peiris, Hiranya V. ;
Weinfurtner, Silke .
JOURNAL OF HIGH ENERGY PHYSICS, 2018, (07)
[5]  
Bu Z., 2021, ARXIV210607830
[6]  
Bu ZQ, 2021, ADV NEUR IN, V34
[7]   CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification [J].
Chen, Chun-Fu ;
Fan, Quanfu ;
Panda, Rameswar .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :347-356
[8]   Visformer: The Vision-friendly Transformer [J].
Chen, Zhengsu ;
Xie, Lingxi ;
Niu, Jianwei ;
Liu, Xuefeng ;
Wei, Longhui ;
Tian, Qi .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :569-578
[9]   ConViT: improving vision transformers with soft convolutional inductive biases [J].
d'Ascoli, Stephane ;
Touvron, Hugo ;
Leavitt, Matthew L. ;
Morcos, Ari S. ;
Biroli, Giulio ;
Sagun, Levent .
JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2022, 2022 (11)
[10]  
De Soham, 2022, ARXIV220413650