Automatic Tuning of Privacy Budgets in Input-Discriminative Local Differential Privacy

被引:4
作者
Murakami, Takao [1 ]
Sei, Yuichi [2 ]
机构
[1] Natl Inst Adv Ind Sci & Technol, Cyber Phys Secur Res Ctr, Tokyo 1350064, Japan
[2] Univ Electrocommun, Grad Sch Informat & Engn, Tokyo 1828585, Japan
基金
日本学术振兴会;
关键词
Privacy; Data privacy; Tuning; Internet of Things; Databases; Proposals; Differential privacy; Frequency estimation; Internet of Things (IoT); local differential privacy (LDP); privacy budget; reidentification; CONFIDENCE-INTERVALS; PROPORTION;
D O I
10.1109/JIOT.2023.3267082
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Local differential privacy (LDP) and its variants have been recently studied to analyze personal data collected from Internet of Things (IoT) devices while strongly protecting user privacy. In particular, a recent study proposes a general privacy notion called input-discriminative LDP (ID-LDP), which introduces a privacy budget for each input value to deal with different levels of sensitivity. However, it is unclear how to set an appropriate privacy budget for each input value, especially, in current situations where reidentification is considered a major risk, e.g., in GDPR. Moreover, the possible number of input values can be very large in IoT. Consequently, it is also extremely difficult to manually check whether a privacy budget for each input value is appropriate. In this article, we propose algorithms to automatically tune privacy budgets in ID-LDP so that obfuscated data strongly prevent reidentification. We also propose a new instance of ID-LDP called one-budget ID-LDP (OneID-LDP) to prevent reidentification with high utility. Through comprehensive experiments using four real data sets, we show that existing instances of ID-LDP lack either utility or privacy-they overprotect personal data or are vulnerable to reidentification attacks. Then, we show that our OneID-LDP mechanisms with our privacy budget tuning algorithm provide much higher utility than LDP mechanisms while strongly preventing reidentification.
引用
收藏
页码:15990 / 16005
页数:16
相关论文
共 64 条
  • [1] Acharya J., 2020, P MACHINE LEARNING R, P52
  • [2] Andr ~es M. E., 2013, P 2013 ACM SIGSAC C, P901
  • [3] [Anonymous], 2020, securitymagazine
  • [4] Local Differential Privacy for Deep Learning
    Arachchige, Pathum Chamikara Mahawaga
    Bertok, Peter
    Khalil, Ibrahim
    Liu, Dongxi
    Camtepe, Seyit
    Atiquzzaman, Mohammed
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (07): : 5827 - 5842
  • [5] The Privacy Blanket of the Shuffle Model
    Balle, Borja
    Bell, James
    Gascon, Adria
    Nissim, Kobbi
    [J]. ADVANCES IN CRYPTOLOGY - CRYPTO 2019, PT II, 2019, 11693 : 638 - 667
  • [6] Local, Private, Efficient Protocols for Succinct Histograms
    Bassily, Raef
    Smith, Adam
    [J]. STOC'15: PROCEEDINGS OF THE 2015 ACM SYMPOSIUM ON THEORY OF COMPUTING, 2015, : 127 - 135
  • [7] Plausible Deniability for Privacy-Preserving Data Synthesis
    Bindschaedler, Vincent
    Shokri, Reza
    Gunter, Carl A.
    [J]. PROCEEDINGS OF THE VLDB ENDOWMENT, 2017, 10 (05): : 481 - 492
  • [8] Interval estimation for a binomial proportion - Comment - Rejoinder
    Brown, LD
    Cai, TT
    DasGupta, A
    Agresti, A
    Coull, BA
    Casella, G
    Corcoran, C
    Mehta, C
    Ghosh, M
    Santner, TJ
    Brown, LD
    Cai, TT
    DasGupta, A
    [J]. STATISTICAL SCIENCE, 2001, 16 (02) : 101 - 133
  • [9] Chatzikokolakis Konstantinos, 2013, Privacy Enhancing Technologies.13th International Symposium, PETS 2013. Proceedings: LNCS 7981, P82, DOI 10.1007/978-3-642-39077-7_5
  • [10] Chelmis C, 2015, PROCEEDINGS 2015 IEEE INTERNATIONAL CONFERENCE ON BIG DATA, P2223, DOI 10.1109/BigData.2015.7364011