Fast Yet Effective Machine Unlearning

被引:31
作者
Tarun, Ayush K. [1 ]
Chundawat, Vikram S. [1 ]
Mandal, Murari [2 ,3 ]
Kankanhalli, Mohan [4 ]
机构
[1] Mavvex Labs, Faridabad 121001, India
[2] Natl Univ Singapore, Sch Comp, Singapore 117417, Singapore
[3] Kalinga Inst Ind Technol KIIT, Sch Comp Engn, Bhubaneswar 751024, India
[4] Natl Univ Singapore NUS, Sch Comp, Singapore 117417, Singapore
基金
新加坡国家研究基金会;
关键词
Data models; Training; Data privacy; Deep learning; Task analysis; Privacy; Training data; forgetting; machine unlearning; privacy in artificial intelligence (AI);
D O I
10.1109/TNNLS.2023.3266233
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unlearning the data observed during the training of a machine learning (ML) model is an important task that can play a pivotal role in fortifying the privacy and security of ML-based applications. This article raises the following questions: 1) can we unlearn a single or multiple class(es) of data from an ML model without looking at the full training data even once? and 2) can we make the process of unlearning fast and scalable to large datasets, and generalize it to different deep networks? We introduce a novel machine unlearning framework with error-maximizing noise generation and impair-repair based weight manipulation that offers an efficient solution to the above questions. An error-maximizing noise matrix is learned for the class to be unlearned using the original model. The noise matrix is used to manipulate the model weights to unlearn the targeted class of data. We introduce impair and repair steps for a controlled manipulation of the network weights. In the impair step, the noise matrix along with a very high learning rate is used to induce sharp unlearning in the model. Thereafter, the repair step is used to regain the overall performance. With very few update steps, we show excellent unlearning while substantially retaining the overall model accuracy. Unlearning multiple classes requires a similar number of update steps as for a single class, making our approach scalable to large problems. Our method is quite efficient in comparison to the existing methods, works for multiclass unlearning, does not put any constraints on the original optimization mechanism or network design, and works well in both small and large-scale vision tasks. This work is an important step toward fast and easy implementation of unlearning in deep networks. Source code: https://github.com/vikram2000b/Fast-Machine-Unlearning.
引用
收藏
页码:13046 / 13055
页数:10
相关论文
共 53 条
  • [1] Deep Learning with Differential Privacy
    Abadi, Martin
    Chu, Andy
    Goodfellow, Ian
    McMahan, H. Brendan
    Mironov, Ilya
    Talwar, Kunal
    Zhang, Li
    [J]. CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, : 308 - 318
  • [2] Achille A., 2019, ARXIV
  • [3] Machine unlearning: linear filtration for logit-based classifiers
    Baumhauer, Thomas
    Schoettle, Pascal
    Zeppelzauer, Matthias
    [J]. MACHINE LEARNING, 2022, 111 (09) : 3203 - 3226
  • [4] Bourtoule L, 2021, P IEEE S SECUR PRIV, P141, DOI 10.1109/SP40001.2021.00019
  • [5] Brophy J., 2021, INT C MACHINE LEARNI, P1092
  • [6] VGGFace2: A dataset for recognising faces across pose and age
    Cao, Qiong
    Shen, Li
    Xie, Weidi
    Parkhi, Omkar M.
    Zisserman, Andrew
    [J]. PROCEEDINGS 2018 13TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION (FG 2018), 2018, : 67 - 74
  • [7] Towards Making Systems Forget with Machine Unlearning
    Cao, Yinzhi
    Yang, Junfeng
    [J]. 2015 IEEE SYMPOSIUM ON SECURITY AND PRIVACY SP 2015, 2015, : 463 - 480
  • [8] Recommendation Unlearning
    Chen, Chong
    Sun, Fei
    Zhang, Min
    Ding, Bolin
    [J]. PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, : 2768 - 2777
  • [9] When Machine Unlearning Jeopardizes Privacy
    Chen, Min
    Zhang, Zhikun
    Wang, Tianhao
    Backes, Michael
    Humbert, Mathias
    Zhang, Yang
    [J]. CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2021, : 896 - 911
  • [10] Choromanska A, 2015, JMLR WORKSH CONF PRO, V38, P192