Distributional Black-Box Model Inversion Attack With Multi-Agent Reinforcement Learning

被引:0
作者
Bao, Huan [1 ,2 ]
Wei, Kaimin [1 ,2 ]
Wu, Yongdong [1 ,2 ]
Qian, Jin [1 ,2 ]
Deng, Robert H. [3 ]
机构
[1] Jinan Univ, Coll Cyber Secur, Guangdong Prov Key Lab Data Secur & Privacy Protec, Guangzhou 510632, Peoples R China
[2] Jinan Univ, Guangdong Inst Smart Educ, Guangzhou 510632, Peoples R China
[3] Singapore Management Univ, Sch Informat Syst, Singapore 178902, Singapore
基金
中国国家自然科学基金;
关键词
Generative adversarial networks; Closed box; Training; Codes; Data models; Data privacy; Glass box; Image reconstruction; Training data; Probability distribution; Distributional model inversion (MI) attack; deep learning; multi-agent reinforcement learning (MARL); black-box attack; ROBUSTNESS;
D O I
10.1109/TIFS.2025.3564043
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Model Inversion (MI) attacks based on Generative Adversarial Networks (GAN) aim to recover private training data from complex deep learning models by searching codes in the latent space. However, this method merely searches in a deterministic latent space, resulting in suboptimal latent codes. Additionally, existing distributional MI schemes assume that an attacker can access the structures and parameters of the target model, which is not always feasible in practice. To address these limitations, this paper proposes a novel Distributional Black-Box Model Inversion (DBB-MI) attack by constructing a probabilistic latent space for searching private data. Specifically, DBB-MI does not require the target model's parameters or specialized GAN training. Instead, it identifies the latent probability distribution by integrating the output of the target model with multi-agent reinforcement learning techniques. Then, it randomly selects latent codes from the latent probability distribution to uncover private data. As the latent probability distribution closely mirrors the target privacy data in the latent space, the recovered data effectively leaks the privacy of the target model's training samples. Extensive experiments conducted on diverse datasets and networks demonstrate that our DBB-MI outperforms state-of-the-art MI attacks in terms of attack accuracy, K-nearest neighbor feature distance, and peak signal-to-noise ratio.
引用
收藏
页码:5425 / 5437
页数:13
相关论文
共 38 条
[1]  
An S., 2022, P NETW DISTR SYST SE
[2]   Killing Two Birds with One Stone: Efficient and Robust Training of Face Recognition CNNs by Partial FC [J].
An, Xiang ;
Deng, Jiankang ;
Guo, Jia ;
Feng, Ziyong ;
Zhu, XuHan ;
Yang, Jing ;
Liu, Tongliang .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :4032-4041
[3]  
Bao H., 2024, P 35 BRIT MACH VIS C, P1
[4]   FePN: A robust feature purification network to defend against adversarial examples [J].
Cao, Dongliang ;
Wei, Kaimin ;
Wu, Yongdong ;
Zhang, Jilian ;
Feng, Bingwen ;
Chen, Jinpeng .
COMPUTERS & SECURITY, 2023, 134
[5]   Knowledge-Enriched Distributional Model Inversion Attacks [J].
Chen, Si ;
Kahla, Mostafa ;
Jia, Ruoxi ;
Qi, Guo-Jun .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :16158-16167
[6]  
Chen YD, 2022, ADV NEUR IN
[7]   Know You at One Glance: A Compact Vector Representation for Low-Shot Learning [J].
Cheng, Yu ;
Zhao, Jian ;
Wang, Zhecan ;
Xu, Yan ;
Jayashree, Karlekar ;
Shen, Shengmei ;
Feng, Jiashi .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2017), 2017, :1924-1932
[8]   Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures [J].
Fredrikson, Matt ;
Jha, Somesh ;
Ristenpart, Thomas .
CCS'15: PROCEEDINGS OF THE 22ND ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2015, :1322-1333
[9]  
Fredrikson M, 2014, PROCEEDINGS OF THE 23RD USENIX SECURITY SYMPOSIUM, P17
[10]  
Gupta Jayesh K., 2017, Autonomous Agents and Multiagent Systems, AAMAS 2017: Workshops, Best Papers. Revised Selected Papers: LNAI 10642, P66, DOI 10.1007/978-3-319-71682-4_5