Privacy Re-Identification Attacks on Tabular GANs

被引:1
作者
Alshantti, Abdallah [1 ,2 ]
Rasheed, Adil [1 ]
Westad, Frank [1 ]
机构
[1] Norwegian Univ Sci & Technol, Dept Engn Cybernet, N-7034 Trondheim, Norway
[2] Dept Customer Risk Rating, DNB ASA, Oslo, Norway
来源
SECURITY AND PRIVACY | 2025年 / 8卷 / 01期
关键词
generative models; privacy risk; reconstruction attacks; tabular data; MULTIOBJECTIVE EVOLUTIONARY ALGORITHMS; NSGA-II;
D O I
10.1002/spy2.469
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Generative models are effective in producing realistic tabular synthetic data that resembles the properties and distribution of real datasets. While synthetic data has numerous applications across various domains, generative models are susceptible to overfitting, which can lead to the leakage of sensitive information from training data. Privacy attacks exacerbate this issue by attempting to identify original data records from synthetic data, especially when the attacker possesses some knowledge about the generative model. In this work, we investigate the privacy risks associated with using generative adversarial networks (GANs) to create tabular synthetic datasets. More specifically, we develop privacy reconstruction attacks designed to identify training samples by minimizing their proximity to synthetic records. Our experimental analysis considers various scenarios of reconstruction attacks, in which attackers have different levels of access to the generative models. Additionally, we propose multi-objective optimization using evolutionary algorithms to perturb synthetic samples closer to original training data points. The experimental results show that reconstruction attacks can effectively identify training samples, with privacy threats significantly increasing when attackers have access to the generative model. Furthermore, our findings indicate that using evolutionary algorithms in reconstruction attacks further heightens the risk of identifying confidential samples. Comparing our attacks against state-of-the-art privacy attacks on tabular GANs further reveals that our reconstructions attacks are considerably more effective in recovering real data records.
引用
收藏
页数:18
相关论文
共 85 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]  
Agrawal R, 2000, SIGMOD REC, V29, P439, DOI 10.1145/335191.335438
[3]   CasTGAN: Cascaded Generative Adversarial Network for Realistic Tabular Data Synthesis [J].
Alshantti, Abdallah ;
Varagnolo, Damiano ;
Rasheed, Adil ;
Rahmati, Aria ;
Westad, Frank .
IEEE ACCESS, 2024, 12 :13213-13232
[4]  
Avodji U., 2019, GAMIN ADVERSARIAL AP
[5]   Generative Adversarial Networks: A Survey Toward Private and Secure Applications [J].
Cai, Zhipeng ;
Xiong, Zuobin ;
Xu, Honghui ;
Wang, Peng ;
Li, Wei ;
Pan, Yi .
ACM COMPUTING SURVEYS, 2021, 54 (06)
[6]  
Carlini N, 2022, P IEEE S SECUR PRIV, P1897, DOI [10.1109/SP46214.2022.9833649, 10.1109/SP46214.2022.00090]
[7]   GAN-Leaks: A Taxonomy of Membership Inference Attacks against Generative Models [J].
Chen, Dingfan ;
Yu, Ning ;
Zhang, Yang ;
Fritz, Mario .
CCS '20: PROCEEDINGS OF THE 2020 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2020, :343-362
[8]  
Chen Dingfan, 2020, Advances in Neural Information Processing Systems, V33
[9]   PAR-GAN: Improving the Generalization of Generative Adversarial Networks Against Membership Inference Attacks [J].
Chen, Junjie ;
Wang, Wendy Hui ;
Gao, Hongchang ;
Shi, Xinghua .
KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, :127-137
[10]  
Choi E., 2017, PMLR, P286, DOI DOI 10.48550/ARXIV.1703.06490