Sharper Utility Bounds for Differentially Private Models: Smooth and Non-smooth

被引:0
作者
Kang, Yilin [1 ,2 ]
Liu, Yong [3 ]
Li, Jian [1 ]
Wang, Weiping [1 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Renmin Univ China, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022 | 2022年
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
machine learning; differential privacy; excess population risk;
D O I
10.1145/3511808.3557451
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, by introducing Generalized Bernstein condition, we propose the first O(root p/n is an element of) high probability excess population risk bound for differentially private algorithms under the assumptions G-Lipschitz, L-smooth, and Polyak-Lojasiewicz condition, based on gradient perturbation method. If we replace the propertiesG-Lipschitz and L-smooth by alpha-Holder smoothness (which can be used in non-smooth setting), the high probability bound comes to O(n-alpha/1+2 alpha) w.r.t. n, which cannot achieve O (1/n) when alpha is an element of (0, 1]. To solve this problem, we propose a variant of gradient perturbation method, max{1,g}-Normalized Gradient Perturbation (m-NGP). We further show that by normalization, the high probability excess population risk bound under assumptions alpha-Holder smooth and Polyak-Lojasiewicz condition can achieve O(root p/n is an element of), which is the first O (1/n) high probability excess population risk bound w.r.t. n for differentially private algorithms under non-smooth conditions. Moreover, experimental results show that m-NGP improves the performance of the differentially private model over real datasets.
引用
收藏
页码:951 / 961
页数:11
相关论文
共 63 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]  
Backes Michael, 2016, P 2016 ACM SIGSAC C, P319, DOI 10.1145/
[3]   STATISTICAL GUARANTEES FOR THE EM ALGORITHM: FROM POPULATION TO SAMPLE-BASED ANALYSIS [J].
Balakrishnan, Sivaraman ;
Wainwrightt, Martin J. ;
Yu, Bin .
ANNALS OF STATISTICS, 2017, 45 (01) :77-120
[4]  
Bartlett P. L., 2002, Computational Learning Theory. 15th Annual Conference on Computational Learning Theory, COLT 2002. Proceedings (Lecture Notes in Artificial Intelligence Vol.2375), P44
[5]   Convexity, classification, and risk bounds [J].
Bartlett, PL ;
Jordan, MI ;
McAuliffe, JD .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2006, 101 (473) :138-156
[6]  
Bassily R, 2019, ADV NEUR IN, V32
[7]   Private Empirical Risk Minimization: Efficient Algorithms and Tight Error Bounds [J].
Bassily, Raef ;
Smith, Adam ;
Thakurta, Abhradeep .
2014 55TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS 2014), 2014, :464-473
[8]  
Bassily Raef, 2020, ADV NEURAL INFORM PR, V33
[9]  
Bernstein Garrett, 2019, ADV NEURAL INFORM PR, V32, P525
[10]  
Bontempi Gianluca, 2018, ULB The Machine Learning Group