Label-Only Membership Inference Attack Based on Model Explanation

被引:0
作者
Ma, Yao [1 ]
Zhai, Xurong [1 ]
Yu, Dan [1 ]
Yang, Yuli [1 ]
Wei, Xingyu [2 ]
Chen, Yongle [1 ]
机构
[1] Taiyuan Univ Technol, Coll Comp Sci & Technol, Jinzhong 030600, Peoples R China
[2] Tsinghua Univ, Res Ctr Identificat & Resolut Syst, Jiashan Novat Ctr, Yangtze Delta Reg Inst, Beijing 314100, Zhejiang, Peoples R China
关键词
Machine Learning; Membership Inference Attack; Forgettable Examples; Feature Attribution; Confidence Estimate;
D O I
10.1007/s11063-024-11682-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is well known that machine learning models (e.g., image recognition) can unintentionally leak information about the training set. Conventional membership inference relies on posterior vectors, and this task becomes extremely difficult when the posterior is masked. However, current label-only membership inference attacks require a large number of queries during the generation of adversarial samples, and thus incorrect inference generates a large number of invalid queries. Therefore, we introduce a label-only membership inference attack based on model explanations. It can transform a label-only attack into a traditional membership inference attack by observing neighborhood consistency and perform fine-grained membership inference for vulnerable samples. We use feature attribution to simplify the high-dimensional neighborhood sampling process, quickly identify decision boundaries and recover a posteriori vectors. It also compares different privacy risks faced by different samples through finding vulnerable samples. The method is validated on CIFAR-10, CIFAR-100 and MNIST datasets. The results show that membership attributes can be identified even using a simple sampling method. Furthermore, vulnerable samples expose the model to greater privacy risks.
引用
收藏
页数:17
相关论文
共 28 条
  • [1] Bengio Yoshua, 2019, ICLR
  • [2] Carlini N, 2022, P IEEE S SECUR PRIV, P1897, DOI [10.1109/SP46214.2022.00090, 10.1109/SP46214.2022.9833649]
  • [3] Practical Membership Inference Attack Against Collaborative Inference in Industrial IoT
    Chen, Hanxiao
    Li, Hongwei
    Dong, Guishan
    Hao, Meng
    Xu, Guowen
    Huang, Xiaoming
    Liu, Zhe
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (01) : 477 - 487
  • [4] Choquette-Choo CA, 2021, PR MACH LEARN RES, V139
  • [5] Conti M., 2022, P 15 ACM WORKSH ART, P1
  • [6] CS-MIA: Membership inference attack based on prediction confidence series in federated learning
    Gu, Yuhao
    Bai, Yuebin
    Xu, Shubin
    [J]. JOURNAL OF INFORMATION SECURITY AND APPLICATIONS, 2022, 67
  • [7] MemGuard: Defending against Black-Box Membership Inference Attacks via Adversarial Examples
    Jia, Jinyuan
    Salem, Ahmed
    Backes, Michael
    Zhang, Yang
    Gong, Neil Zhenqiang
    [J]. PROCEEDINGS OF THE 2019 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'19), 2019, : 259 - 274
  • [8] Leino K, 2020, PROCEEDINGS OF THE 29TH USENIX SECURITY SYMPOSIUM, P1605
  • [9] Membership Leakage in Label-Only Exposures
    Li, Zheng
    Zhang, Yang
    [J]. CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2021, : 880 - 895
  • [10] Liu Yiyong, 2022, P 2022 ACM SIGSAC C, P2085