Zero-Shot Machine Unlearning

被引:46
作者
Chundawat, Vikram S. [1 ]
Tarun, Ayush K. [1 ]
Mandal, Murari [1 ,2 ]
Kankanhalli, Mohan [1 ]
机构
[1] Natl Univ Singapore, Sch Comput ing, Singapore 117417, Singapore
[2] Kalinga Inst Ind Technol KIIT, Sch Comp Engn, Bhubaneswar 751024, India
基金
新加坡国家研究基金会;
关键词
Data models; Training; Data privacy; Training data; Computational modeling; Regulation; Machine learning; Machine unlearning; machine learning security and privacy; data privacy;
D O I
10.1109/TIFS.2023.3265506
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Modern privacy regulations grant citizens the right to be forgotten by products, services and companies. In case of machine learning (ML) applications, this necessitates deletion of data not only from storage archives but also from ML models. Due to an increasing need for regulatory compliance required for ML applications, machine unlearning is becoming an emerging research problem. The right to be forgotten requests come in the form of removal of a certain set or class of data from the already trained ML model. Practical considerations preclude retraining of the model from scratch after discarding the deleted data. The few existing studies use either the whole training data, or a subset of training data, or some metadata stored during training to update the model weights for unlearning. However, strict regulatory compliance requires time-bound deletion of data. Thus, in many cases, no data related to the training process or training samples may be accessible even for the unlearning purpose. We therefore ask the question: is it possible to achieve unlearning with zero training samples? In this paper, we introduce the novel problem of zero-shot machine unlearning that caters for the extreme but practical scenario where zero original data samples are available for use. We then propose two novel solutions for zero-shot machine unlearning based on (a) error minimizing-maximizing noise and (b) gated knowledge transfer. These methods remove the information of the forget data from the model while maintaining the model efficacy on the retain data. The zero-shot approach offers good protection against the model inversion attacks and membership inference attacks. We introduce a new evaluation metric, Anamnesis Index (AIN) to effectively measure the quality of the unlearning method. The experiments show promising results for unlearning in deep learning models on benchmark vision data-sets. The source code is available here: https://github.com/ayu987/zero-shot-unlearning
引用
收藏
页码:2345 / 2354
页数:10
相关论文
共 45 条
[1]   Machine Unlearning [J].
Bourtoule, Lucas ;
Chandrasekaran, Varun ;
Choquette-Choo, Christopher A. ;
Jia, Hengrui ;
Travers, Adelin ;
Zhang, Baiwu ;
Lie, David ;
Papernot, Nicolas .
2021 IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP, 2021, :141-159
[2]  
Brophy J., 2021, P MACHINE LEARNING R, P1092
[3]   Towards Making Systems Forget with Machine Unlearning [J].
Cao, Yinzhi ;
Yang, Junfeng .
2015 IEEE SYMPOSIUM ON SECURITY AND PRIVACY SP 2015, 2015, :463-480
[4]  
Carlini N, 2019, PROCEEDINGS OF THE 28TH USENIX SECURITY SYMPOSIUM, P267
[5]   When Machine Unlearning Jeopardizes Privacy [J].
Chen, Min ;
Zhang, Zhikun ;
Wang, Tianhao ;
Backes, Michael ;
Humbert, Mathias ;
Zhang, Yang .
CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2021, :896-911
[6]  
Chundawat V. S., 2023, PROC AAAI C ARTIF IN, P1
[7]   Robust Traceability from Trace Amounts [J].
Dwork, Cynthia ;
Smith, Adam ;
Steinke, Thomas ;
Ullman, Jonathan ;
Vadhan, Salil .
2015 IEEE 56TH ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, 2015, :650-669
[8]   Does Learning Require Memorization? A Short Tale about a Long Tail [J].
Feldman, Vitaly .
PROCEEDINGS OF THE 52ND ANNUAL ACM SIGACT SYMPOSIUM ON THEORY OF COMPUTING (STOC '20), 2020, :954-959
[9]   Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures [J].
Fredrikson, Matt ;
Jha, Somesh ;
Ristenpart, Thomas .
CCS'15: PROCEEDINGS OF THE 22ND ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2015, :1322-1333
[10]   Property Inference Attacks on Fully Connected Neural Networks using Permutation Invariant Representations [J].
Ganju, Karan ;
Wang, Qi ;
Yang, Wei ;
Gunter, Carl A. ;
Borisov, Nikita .
PROCEEDINGS OF THE 2018 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'18), 2018, :619-633