Stochastic privacy-preserving methods for nonconvex sparse learning

被引:1
作者
Liang, Guannan [1 ]
Tong, Qianqian [1 ]
Ding, Jiahao [2 ]
Pan, Miao [2 ]
Bi, Jinbo [1 ]
机构
[1] Univ Connecticut, Storrs, CT 06269 USA
[2] Univ Houston, Houston, TX USA
基金
美国国家卫生研究院; 美国国家科学基金会;
关键词
Sparse learning; Differential privacy; Stochastic algorithm; NOISE;
D O I
10.1016/j.ins.2022.09.062
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sparse learning is essential in mining high-dimensional data. Iterative hard thresholding (IHT) methods are effective for optimizing nonconvex objectives for sparse learning. However, IHT methods are vulnerable to adversary attacks that infer sensitive data. Although pioneering works attempted to relieve such vulnerability, they confront the issue of high computational cost for large-scale problems. We propose two differentially private stochastic IHT: one based on the stochastic gradient descent method (DP-SGD-HT) and the other based on the stochastically controlled stochastic gradient method (DP-SCSG-HT). The DP-SGD-HT method perturbs stochastic gradients with small Gaussian noise rather than full gradients, which are computationally expensive. As a result, computational complexity is reduced from O(nlog(n)) to a lower O(blog(n)), where n is the sample size and b is the mini-batch size used to compute stochastic gradients. The DP-SCSG-HT method further perturbs the stochastic gradients controlled by largebatch snapshot gradients to reduce stochastic gradient variance. We prove that both algorithms guarantee differential privacy and have linear convergence rates with estimation bias. A utility analysis examines the relationship between convergence rate and the level of perturbation, yielding the best-known utility bound for nonconvex sparse optimization. Extensive experiments show that our algorithms outperform existing methods.
引用
收藏
页码:567 / 585
页数:19
相关论文
共 48 条
[1]  
A.E.C. Cloud, AMAZON WEB SERVICES
[2]  
Abadi M., 2015, TensorFlow: Large-scale machine Learning on heterogeneous distributed systems, DOI DOI 10.48550/ARXIV.1603.04467
[3]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[4]   Management of validation of HPLC method for determination of acetylsalicylic acid impurities in a new pharmaceutical product [J].
Kowalska, Malgorzata ;
Wozniak, Magdalena ;
Kijek, Michal ;
Mitrosz, Paulina ;
Szakiel, Jerzy ;
Turek, Pawel .
SCIENTIFIC REPORTS, 2022, 12 (01)
[5]  
[Anonymous], 2012, P 25 ANN C LEARNING
[6]  
Babanezhad R, 2015, ADV NEUR IN, V28
[7]  
Bahmani S, 2013, J MACH LEARN RES, V14, P807
[8]   Private Empirical Risk Minimization: Efficient Algorithms and Tight Error Bounds [J].
Bassily, Raef ;
Smith, Adam ;
Thakurta, Abhradeep .
2014 55TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS 2014), 2014, :464-473
[9]   Iterative hard thresholding for compressed sensing [J].
Blumensath, Thomas ;
Davies, Mike E. .
APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2009, 27 (03) :265-274
[10]  
Chaudhuri K, 2011, J MACH LEARN RES, V12, P1069