Privacy Enhancement Via Dummy Points in the Shuffle Model

被引:3
作者
Li, Xiaochen [1 ,2 ]
Liu, Weiran [3 ]
Feng, Hanwen [3 ]
Huang, Kunzhe [1 ]
Hu, Yuke [1 ,2 ]
Liu, Jinfei [1 ]
Ren, Kui [1 ]
Qin, Zhan [1 ,2 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310027, Peoples R China
[2] ZJU Hangzhou Global Sci & Technol Innovat Ctr, Hangzhou 311215, Peoples R China
[3] Alibaba Grp, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Privacy; Protocols; Data models; Histograms; Differential privacy; Analytical models; Estimation error; Dummy points; local differential privacy; privacy enhancement; shuffle; DIFFERENTIAL PRIVACY; PROVABLY-SECURE; EFFICIENT; UTILITY;
D O I
10.1109/TDSC.2023.3263162
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The shuffle model is recently proposed to address the issue of severe utility loss in Local Differential Privacy (LDP) due to distributed data randomization. In the shuffle model, a shuffler is utilized to break the link between the user identity and the message uploaded to the data analyst. Since less noise needs to be introduced to achieve the same privacy guarantee, following this paradigm, the utility of privacy-preserving data collection is improved. We propose DUMP (DUMmy-Point-based), a framework for privacy-preserving histogram estimation in the shuffle model. The core of DUMP is a new concept of dummy blanket, which enables enhancing privacy by just introducing dummy points on the user side and further improving the utility of the shuffle model. We instantiate DUMP by proposing two protocols: pureDUMP and mixDUMP, and conduct a comprehensive experimental evaluation to compare them with existing protocols. The experimental results show that, under the same privacy guarantee, (1) the proposed protocols have significant improvements in communication efficiency over all existing multi-message protocols, by at least 3 orders of magnitude; (2) they achieve competitive utility, while the only known protocol (Ghazi et al. PMLR 2020) having better utility than ours employs hard-to-exactly-sample distributions which are vulnerable to floating-point attacks (CCS 2012).
引用
收藏
页码:1001 / 1016
页数:16
相关论文
共 52 条
[1]  
[Anonymous], 2017, Movie rating dataset
[2]  
[Anonymous], 2021, Secure noise generation
[3]  
[Anonymous], 2017, IPUMS Census dataset
[4]  
Apple Differential Privacy Team, 2017, LEARNING PRIVACY SCA
[5]  
Balcer V., 2018, P INN THEOR COMP SCI
[6]  
Balcer V., 2020, LIPIcs, DOI 10.4230/
[7]  
Balle B, 2019, Arxiv, DOI arXiv:1906.09116
[8]  
Balle B, 2020, PR MACH LEARN RES, V108, P2496
[9]   The Privacy Blanket of the Shuffle Model [J].
Balle, Borja ;
Bell, James ;
Gascon, Adria ;
Nissim, Kobbi .
ADVANCES IN CRYPTOLOGY - CRYPTO 2019, PT II, 2019, 11693 :638-667
[10]   PROCHLO: Strong Privacy for Analytics in the Crowd [J].
Bittau, Andrea ;
Erlingsson, Ulfar ;
Maniatis, Petros ;
Mironov, Ilya ;
Raghunathan, Ananth ;
Lie, David ;
Rudominer, Mitch ;
Kode, Ushasree ;
Tinnes, Julien ;
Seefeld, Bernhard .
PROCEEDINGS OF THE TWENTY-SIXTH ACM SYMPOSIUM ON OPERATING SYSTEMS PRINCIPLES (SOSP '17), 2017, :441-459