Label-Only Membership Inference Attack Based on Model Explanation

被引:0
作者
Ma, Yao [1 ]
Zhai, Xurong [1 ]
Yu, Dan [1 ]
Yang, Yuli [1 ]
Wei, Xingyu [2 ]
Chen, Yongle [1 ]
机构
[1] Taiyuan Univ Technol, Coll Comp Sci & Technol, Jinzhong 030600, Peoples R China
[2] Tsinghua Univ, Res Ctr Identificat & Resolut Syst, Jiashan Novat Ctr, Yangtze Delta Reg Inst, Beijing 314100, Zhejiang, Peoples R China
关键词
Machine Learning; Membership Inference Attack; Forgettable Examples; Feature Attribution; Confidence Estimate;
D O I
10.1007/s11063-024-11682-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is well known that machine learning models (e.g., image recognition) can unintentionally leak information about the training set. Conventional membership inference relies on posterior vectors, and this task becomes extremely difficult when the posterior is masked. However, current label-only membership inference attacks require a large number of queries during the generation of adversarial samples, and thus incorrect inference generates a large number of invalid queries. Therefore, we introduce a label-only membership inference attack based on model explanations. It can transform a label-only attack into a traditional membership inference attack by observing neighborhood consistency and perform fine-grained membership inference for vulnerable samples. We use feature attribution to simplify the high-dimensional neighborhood sampling process, quickly identify decision boundaries and recover a posteriori vectors. It also compares different privacy risks faced by different samples through finding vulnerable samples. The method is validated on CIFAR-10, CIFAR-100 and MNIST datasets. The results show that membership attributes can be identified even using a simple sampling method. Furthermore, vulnerable samples expose the model to greater privacy risks.
引用
收藏
页数:17
相关论文
共 28 条
[1]  
Carlini N, 2022, P IEEE S SECUR PRIV, P1897, DOI [10.1109/SP46214.2022.00090, 10.1109/SP46214.2022.9833649]
[2]   Practical Membership Inference Attack Against Collaborative Inference in Industrial IoT [J].
Chen, Hanxiao ;
Li, Hongwei ;
Dong, Guishan ;
Hao, Meng ;
Xu, Guowen ;
Huang, Xiaoming ;
Liu, Zhe .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (01) :477-487
[3]  
Choquette-Choo CA, 2021, PR MACH LEARN RES, V139
[4]  
Conti M., 2022, P 15 ACM WORKSH ART, P1
[5]  
Gordon Geoffrey J., 2019, ICLR
[6]   CS-MIA: Membership inference attack based on prediction confidence series in federated learning [J].
Gu, Yuhao ;
Bai, Yuebin ;
Xu, Shubin .
JOURNAL OF INFORMATION SECURITY AND APPLICATIONS, 2022, 67
[7]   MemGuard: Defending against Black-Box Membership Inference Attacks via Adversarial Examples [J].
Jia, Jinyuan ;
Salem, Ahmed ;
Backes, Michael ;
Zhang, Yang ;
Gong, Neil Zhenqiang .
PROCEEDINGS OF THE 2019 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'19), 2019, :259-274
[8]  
Leino K, 2020, PROCEEDINGS OF THE 29TH USENIX SECURITY SYMPOSIUM, P1605
[9]   Membership Leakage in Label-Only Exposures [J].
Li, Zheng ;
Zhang, Yang .
CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2021, :880-895
[10]  
Liu Y., 2022, P 2022 ACM SIGS C, P2085, DOI DOI 10.1145/3548606.3560684