MUter: Machine Unlearning on Adversarially Trained Models

被引:4
|
作者
Liu, Junxu [1 ]
Xue, Mingsheng [2 ]
Lou, Jian [3 ,6 ]
Zhang, Xiaoyu [4 ]
Xiong, Li [5 ]
Qin, Zhan [3 ,6 ]
机构
[1] Renmin Univ China, Beijing, Peoples R China
[2] Xidian Univ, Guangzhou Inst Technol, Guangzhou, Peoples R China
[3] Zhejiang Univ, Hangzhou, Peoples R China
[4] Xidian Univ, Xian, Peoples R China
[5] Emory Univ, Atlanta, GA 30322 USA
[6] ZJU Hangzhou Global Sci & Technol Innovat Ctr, Hangzhou, Peoples R China
来源
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV | 2023年
基金
美国国家科学基金会;
关键词
ATTACKS; FORGET;
D O I
10.1109/ICCV51070.2023.00451
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine unlearning is an emerging task of removing the influence of selected training datapoints from a trained model upon data deletion requests, which echoes the widely enforced data regulations mandating the Right to be Forgotten. Many unlearning methods have been proposed recently, achieving significant efficiency gains over the naive baseline of retraining from scratch. However, existing methods focus exclusively on unlearning from standard training models and do not apply to adversarial training models (ATMs) despite their popularity as effective defenses against adversarial examples. During adversarial training, the training data are involved in not only an outer loop for minimizing the training loss, but also an inner loop for generating the adversarial perturbation. Such bi-level optimization greatly complicates the influence measure for the data to be deleted and renders the unlearning more challenging than standard model training with single-level optimization. This paper proposes a new approach called MUter for unlearning from ATMs. We derive a closed-form unlearning step underpinned by a total Hessian-related data influence measure, while existing methods can mis-capture the data influence associated with the indirect Hessian part. We further alleviate the computational cost by introducing a series of approximations and conversions to avoid the most computationally demanding parts of Hessian inversions. The efficiency and effectiveness of MUter have been validated through experiments on four datasets using both linear and neural network models.
引用
收藏
页码:4869 / 4879
页数:11
相关论文
共 17 条
  • [1] Learn to Unlearn: Insights Into Machine Unlearning
    Qu, Youyang
    Yuan, Xin
    Ding, Ming
    Ni, Wei
    Rakotoarivelo, Thierry
    Smith, David
    COMPUTER, 2024, 57 (03) : 79 - 90
  • [2] Towards Unbounded Machine Unlearning
    Kurmanji, Meghdad
    Triantafillou, Peter
    Hayes, Jamie
    Triantafillou, Eleni
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Ensuring User Privacy and Model Security via Machine Unlearning: A Review
    Tang, Yonghao
    Cai, Zhiping
    Liu, Qiang
    Zhou, Tongqing
    Ni, Qiang
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 77 (02): : 2645 - 2656
  • [4] A survey of security and privacy issues of machine unlearning
    Chen, Aobo
    Li, Yangyi
    Zhao, Chenxu
    Huai, Mengdi
    AI MAGAZINE, 2025, 46 (01)
  • [5] Learn What You Want to Unlearn: Unlearning Inversion Attacks against Machine Unlearning
    Hu, Hongsheng
    Wang, Shuo
    Dong, Tian
    Xue, Minhui
    45TH IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP 2024, 2024, : 3257 - 3275
  • [6] Model Sparsity Can Simplify Machine Unlearning
    Jia, Jinghan
    Liu, Jiancheng
    Ram, Parikshit
    Yao, Yuguang
    Liu, Gaowen
    Liu, Yang
    Sharma, Pranay
    Liu, Sijia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] Exploring the Landscape of Machine Unlearning: A Comprehensive Survey and Taxonomy
    Shaik, Thanveer
    Tao, Xiaohui
    Xie, Haoran
    Li, Lin
    Zhu, Xiaofeng
    Li, Qing
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [8] No Matter How You Slice It: Machine Unlearning with SISA Comes at the Expense of Minority Classes
    Koch, Korbinian
    Soll, Marcus
    2023 IEEE CONFERENCE ON SECURE AND TRUSTWORTHY MACHINE LEARNING, SATML, 2023, : 622 - 637
  • [9] A survey on machine unlearning: Techniques and new emerged privacy risks
    Liu, Hengzhu
    Xiong, Ping
    Zhu, Tianqing
    Yu, Philip S.
    JOURNAL OF INFORMATION SECURITY AND APPLICATIONS, 2025, 90
  • [10] Hidden Poison: Machine Unlearning Enables Camouflaged Poisoning Attacks
    Di, Jimmy Z.
    Douglas, Jack
    Acharya, Jayadev
    Kamath, Gautam
    Sekhari, Ayush
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,