Attribute-Based Membership Inference Attacks and Defenses on GANs

被引:1
|
作者
Sun, Hui [1 ]
Zhu, Tianqing [2 ]
Li, Jie [1 ]
Ji, Shoulin [3 ]
Zhou, Wanlei [4 ]
机构
[1] China Univ Geosci, Wuhan 430079, Hubei, Peoples R China
[2] Univ Technol Sydney, Sydney, NSW 2007, Australia
[3] Zhejiang Univ, Hangzhou 310027, Zhejiang, Peoples R China
[4] City Univ Macau, Taipa, Macao, Peoples R China
关键词
Training; Image reconstruction; Generators; Generative adversarial networks; Codes; Privacy; Training data; Membership inference attack; generative adversarial networks; privacy leakage;
D O I
10.1109/TDSC.2023.3305591
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
With breakthroughs in high-resolution image generation, applications for disentangled generative adversarial networks (GANs) have attracted much attention. At the same time, the privacy issues associated with GAN models have been raising many concerns. Membership inference attacks (MIAs), where an adversary attempts to determine whether or not a sample has been used to train the victim model, are a major risk with GANs. In prior research, scholars have shown that successful MIAs can be mounted by leveraging overfit images. However, high-resolution images make the existing MIAs fail due to their complexity. And the nature of disentangled GANs is such that the attributes are overfitting, which means that, for an MIA to be successful, it must likely be based on overfitting attributes. Furthermore, given the empirical difficulties with obtaining independent and identically distributed (IID) candidate samples, choosing the non-trivial attributes of candidate samples as the target for exploring overfitting would be a more preferable choice. Hence, in this article, we propose a series of attribute-based MIAs that considers both black-box and white-box settings. The attacks are performed on the generator, and the inferences are derived by overfitting the non-trivial attributes. Additionally, we put forward a novel perspective on model generalization and a possible defense by evaluating the overfitting status of each individual attribute. A series of empirical evaluations in both settings demonstrate that the attacks remain stable and successful when using non-IID candidate samples. Further experiments illustrate that each attribute exhibits a distinct overfitting status. Moreover, manually generalizing highly overfitting attributes significantly reduces the risk of privacy leaks.
引用
收藏
页码:2376 / 2393
页数:18
相关论文
共 50 条
  • [41] Membership Inference Attacks against GNN-based Hardware Trojan Detection
    Hasegawa, Kento
    Yamashita, Kazuki
    Hidano, Seira
    Fukushima, Kazuhide
    Hashimoto, Kazuo
    Togawa, Nozomu
    2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, 2024, : 1222 - 1229
  • [42] Resisting membership inference attacks through knowledge distillation
    Zheng, Junxiang
    Cao, Yongzhi
    Wang, Hanpin
    NEUROCOMPUTING, 2021, 452 : 114 - 126
  • [43] Investigating Membership Inference Attacks under Data Dependencies
    Humphries, Thomas
    Oya, Simon
    Tulloch, Lindsey
    Rafuse, Matthew
    Goldberg, Ian
    Hengartner, Urs
    Kerschbaum, Florian
    2023 IEEE 36TH COMPUTER SECURITY FOUNDATIONS SYMPOSIUM, CSF, 2023, : 473 - 488
  • [44] Membership Inference Attacks Against Machine Learning Models
    Shokri, Reza
    Stronati, Marco
    Song, Congzheng
    Shmatikov, Vitaly
    2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, : 3 - 18
  • [45] Defending Against Membership Inference Attacks on Beacon Services
    Venkatesaramani, Rajagopal
    Wan, Zhiyu
    Malin, Bradley A.
    Vorobeychik, Yevgeniy
    ACM TRANSACTIONS ON PRIVACY AND SECURITY, 2023, 26 (03)
  • [46] Reducing Model Memorization to Mitigate Membership Inference Attacks
    Sheikhjaberi, Mehrdad
    Alhadidi, Dima
    2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, 2024, : 79 - 88
  • [47] Image and Attribute Based Convolutional Neural Network Inference Attacks in Social Networks
    Mei, Bo
    Xiao, Yinhao
    Li, Ruinian
    Li, Hong
    Cheng, Xiuzhen
    Sun, Yunchuan
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2020, 7 (02): : 869 - 879
  • [48] Membership Inference Attacks With Token-Level Deduplication on Korean Language Models
    Oh, Myung Gyo
    Park, Leo Hyun
    Kim, Jaeuk
    Park, Jaewoo
    Kwon, Taekyoung
    IEEE ACCESS, 2023, 11 : 10207 - 10217
  • [49] Membership Inference Attacks Against Temporally Correlated Data in Deep Reinforcement Learning
    Gomrokchi, Maziar
    Amin, Susan
    Aboutalebi, Hossein
    Wong, Alexander
    Precup, Doina
    IEEE ACCESS, 2023, 11 : 42796 - 42808
  • [50] Efficient Membership Inference Attacks against Federated Learning via Bias Differences
    Zhang, Liwei
    Li, Linghui
    Li, Xiaoyong
    Cai, Binsi
    Gao, Yali
    Dou, Ruobin
    Chen, Luying
    PROCEEDINGS OF THE 26TH INTERNATIONAL SYMPOSIUM ON RESEARCH IN ATTACKS, INTRUSIONS AND DEFENSES, RAID 2023, 2023, : 222 - 235