Preserving Data Utility in Differentially Private Smart Home Data

被引:0
|
作者
Stirapongsasuti, Sopicha [1 ]
Tiausas, Francis Jerome [1 ]
Nakamura, Yugo [2 ]
Yasumoto, Keiichi [1 ,3 ]
机构
[1] Nara Inst Sci & Technol, Ikoma, Nara 6300192, Japan
[2] Kyushu Univ, Dept Informat Sci & Elect Engn, Fukuoka 8190395, Japan
[3] RIKEN, Ctr Adv Intelligence Project AIP, Tokyo 1030027, Japan
关键词
Differential privacy; machine learning; privacy; smart home; PRESERVATION; EFFICIENT; SYSTEM; CARE;
D O I
10.1109/ACCESS.2024.3390039
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of smart sensors and appliances can provide a lot of services. Nevertheless, the act of aggregating data containing sensitive information related to privacy in a single location poses significant issues. Such information can be misused by a malicious attacker. Also, some previous studies attempted to apply privacy mechanisms, but they decreased data utility. In this paper, we propose privacy protection mechanisms to preserve privacy-sensitive sensor data generated in a smart home. We leverage R & eacute;nyi differential privacy (RDP) to preserve privacy. However, the preliminary result showed that using only RDP still significantly decreases the utility of data. Thus, a novel scheme called feature merging anonymization (FMA) is proposed to preserve privacy while maintaining data utility by merging feature dataframes of the same activities from other homes. Also, the expected trade-off is defined so that data utility should be greater than the privacy preserved. To evaluate the proposed techniques, we define privacy preservation and data utility as inverse accuracy of person identification (PI) and accuracy of activity recognition (AR), respectively. We trained the AR and PI models for two cases with and without FMA, using 2 smart-home open datasets i.e. the HIS and Toyota dataset. As a result, we could lower the accuracy of PI in the HIS and Toyota dataset to 73.85% and 41.18% with FMA respectively compared to 100% without FMA, while maintaining the accuracy of AR at 94.62% and 87.3% with FMA compared to 98.58% and 89.28% without FMA in the HIS and Toyota dataset, respectively. Another experiment was conducted to explore the feasibility of implementing FMA in a local server by partially merging frames of the original activity with frames of other activities at different merging ratios. The results show that the local server can still satisfy the expected trade-off at some ratios.
引用
收藏
页码:56571 / 56581
页数:11
相关论文
共 50 条
  • [41] Differentially Private Distance Learning in Categorical Data
    Elena Battaglia
    Simone Celano
    Ruggero G. Pensa
    Data Mining and Knowledge Discovery, 2021, 35 : 2050 - 2088
  • [42] Publishing Differentially Private Medical Events Data
    Shaked, Sigal
    Rokach, Lior
    AVAILABILITY, RELIABILITY, AND SECURITY IN INFORMATION SYSTEMS, CD-ARES 2016, PAML 2016, 2016, 9817 : 219 - 235
  • [43] A Differentially Private Method for Crowdsourcing Data Submission
    Zhang, Lefeng
    Xiong, Ping
    Zhu, Tianqing
    TRENDS AND APPLICATIONS IN KNOWLEDGE DISCOVERY AND DATA MINING: PAKDD 2018 WORKSHOPS, 2018, 11154 : 142 - 148
  • [44] PrivPfC: differentially private data publication for classification
    Dong Su
    Jianneng Cao
    Ninghui Li
    Min Lyu
    The VLDB Journal, 2018, 27 : 201 - 223
  • [45] Differentially private response mechanisms on categorical data
    Holohan, Naoise
    Leith, Douglas J.
    Mason, Oliver
    DISCRETE APPLIED MATHEMATICS, 2016, 211 : 86 - 98
  • [46] Differentially Private Publication Scheme for Trajectory Data
    Li, Meng
    Zhu, Liehuang
    Zhang, Zijian
    Xu, Rixin
    2016 IEEE FIRST INTERNATIONAL CONFERENCE ON DATA SCIENCE IN CYBERSPACE (DSC 2016), 2016, : 596 - 601
  • [47] Designing Contracts for Trading Private and Heterogeneous Data Using a Biased Differentially Private Algorithm
    Khalili, Mohammad Mahdi
    Zhang, Xueru
    Liu, Mingyan
    IEEE ACCESS, 2021, 9 : 70732 - 70745
  • [48] A differentially private algorithm for location data release
    Ping Xiong
    Tianqing Zhu
    Wenjia Niu
    Gang Li
    Knowledge and Information Systems, 2016, 47 : 647 - 669
  • [49] Differentially Private Data Releasing for Smooth Queries
    Wang, Ziteng
    Jin, Chi
    Fan, Kai
    Zhang, Jiaqi
    Huang, Junliang
    Zhong, Yiqiao
    Wang, Liwei
    JOURNAL OF MACHINE LEARNING RESEARCH, 2016, 17 : 1 - 42
  • [50] Algorithmically Effective Differentially Private Synthetic Data
    He, Yiyun
    Vershynin, Roman
    Zhu, Yizhe
    THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195