Secure Multi-party Computation of Differentially Private Heavy Hitters

被引:28
作者
Boehler, Jonas [1 ]
Kerschbaum, Florian [2 ]
机构
[1] SAP Secur Res, Karlsruhe, Germany
[2] Univ Waterloo, Waterloo, ON, Canada
来源
CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY | 2021年
关键词
Heavy Hitters; Differential Privacy; Secure Multi-party Computation; Sketches; NOISE;
D O I
10.1145/3460120.3484557
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Private learning of top-k, i.e., the k most frequent values also called heavy hitters, is a common industry scenario: Companies want to privately learn, e.g., frequently typed new words to improve suggestions on mobile devices, often used browser settings, telemetry data of frequent crashes, heavily shared articles, etc. Real-world deployments often use local differential privacy, where distributed users share locally randomized data with an untrusted server. Central differential privacy, on the other hand, assumes access to the raw data and applies the randomization only once, on the aggregated result. These solutions either require large amounts of users for high accuracy (local model) or a trusted third party (central model). We present multi-party computation protocols HH and PEM of sketches (succinct data structures) to efficiently compute differentially private top-k: HH has running time linear in the size of the data and is applicable for very small data sets (hundreds of values), and PEM is sublinear in the data domain and provides better accuracy than HH for large data sizes. Our approaches are efficient (practical running time, requiring no output reconstruction as other sketches) and more accurate than local differential privacy even for a small number of users. In our experiments, we were able to securely compute differentially private top-k in less than 11 minutes using 3 semi-honest computation parties distributed over the Internet with inputs from hundreds of users (HH) and input size that is independent of the user count (PEM, excluding count aggregation).
引用
收藏
页码:2361 / 2377
页数:17
相关论文
共 81 条
[31]  
Dong Jinshuo, 2020, PR MACH LEARN RES, P2597
[32]  
Durfee David, 2019, ADV NEURAL INFORM PR
[33]  
Durfee David, 2019, ARXIV190504273
[34]  
Dwork C, 2006, LECT NOTES COMPUT SC, V4004, P486
[35]   Calibrating noise to sensitivity in private data analysis [J].
Dwork, Cynthia ;
McSherry, Frank ;
Nissim, Kobbi ;
Smith, Adam .
THEORY OF CRYPTOGRAPHY, PROCEEDINGS, 2006, 3876 :265-284
[36]   The Algorithmic Foundations of Differential Privacy [J].
Dwork, Cynthia ;
Roth, Aaron .
FOUNDATIONS AND TRENDS IN THEORETICAL COMPUTER SCIENCE, 2013, 9 (3-4) :211-406
[37]  
Dwork Cynthia, 2015, ARXIV151103803
[38]  
Dwork Cynthia, 2006, Int. colloq, autom., lang., P1
[39]  
Eigner Fabienne, 2014, P ANN COMP SEC APPL
[40]   RAPPOR: Randomized Aggregatable Privacy-Preserving Ordinal Response [J].
Erlingsson, Ulfar ;
Pihur, Vasyl ;
Korolova, Aleksandra .
CCS'14: PROCEEDINGS OF THE 21ST ACM CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2014, :1054-1067