DP-Auditorium: A Large-Scale Library for Auditing Differential Privacy

被引:1
作者
Kong, William [1 ]
Medina, Andres Munoz [1 ]
Ribero, Monica [1 ]
Syed, Umar [1 ]
机构
[1] Google Res, New York, NY 10282 USA
来源
45TH IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP 2024 | 2024年
关键词
DIVERGENCE;
D O I
10.1109/SP54263.2024.00195
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
New regulations and increased awareness of data privacy have led to the deployment of new and more efficient differentially private mechanisms across both public institutions and industries. With the growing adoption of differential privacy, there is also a risk of introducing bugs into both the derivation of new mechanisms and their implementation. Ensuring these mechanisms is therefore crucial to ensure proper protection of data. However since differential privacy is not a property of a single output of a mechanism but a property of the mechanism itself, testing whether a mechanism is differentially private is not a trivial task. While ad hoc testing techniques exist under specific assumptions, no concerted effort has been made by the research community to develop a flexible and extendable tool for testing differentially private mechanisms. This paper introduces DP- AUDITORIUM as a step advancing research in this direction. The main idea behind DP-AUDITORIUM is to abstract the problem of testing differential privacy into two steps: (1) measuring the distance between distributions, and (2) finding neighboring datasets where a mechanism generates output distributions maximizing such distance. From a technical point of view, we propose three new algorithms for evaluating the distance between distributions. While these algorithms are well-known in the statistics community, we provide new estimation guarantees by leveraging the fact that we are only interested in verifying whether a mechanism is differentially private, and not on obtaining an exact estimate of the distance between two distributions. DP-AUDITORIUM is easily extensible, as demonstrated in this paper by implementing a well-known approximate differential privacy testing algorithm to our library. Finally, we provide an extensive comparison to date of multiple testers across varying sample sizes and differential privacy parameters, demonstrating that there is no single tester that dominates all others, and that in order to ensure proper testing of mechanisms, one requires a combination of different techniques.
引用
收藏
页码:110 / 126
页数:17
相关论文
共 38 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]  
Andrew G, 2024, Arxiv, DOI arXiv:2302.03098
[3]  
Balle B, 2022, P IEEE S SECUR PRIV, P1138, DOI [10.1109/SP46214.2022.9833677, 10.1109/SP46214.2022.00127]
[4]  
Barthe G, 2013, LECT NOTES COMPUT SC, V7966, P49, DOI 10.1007/978-3-642-39212-2_8
[5]  
Bichsel B., 2021, S SEC PRIV SP
[6]   DP-Sniper: Black-Box Discovery of Differential Privacy Violations using Classifiers [J].
Bichsel, Benjamin ;
Steffen, Samuel ;
Bogunovic, Ilija ;
Vechev, Martin .
2021 IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP, 2021, :391-409
[7]   Variational Representations and Neural Network Estimation of Renyi Divergences [J].
Birrell, Jeremiah ;
Dupuis, Paul ;
Katsoulakis, Markos A. ;
Rey-Bellet, Luc ;
Wang, Jie .
SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2021, 3 (04) :1093-1116
[8]  
Canonne CL, 2020, ADV NEUR IN, V33
[9]   GAN-Leaks: A Taxonomy of Membership Inference Attacks against Generative Models [J].
Chen, Dingfan ;
Yu, Ning ;
Zhang, Yang ;
Fritz, Mario .
CCS '20: PROCEEDINGS OF THE 2020 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2020, :343-362
[10]  
Desautels T., 2014, Journal of Machine Learning Research