FedRecovery: Differentially Private Machine Unlearning for Federated Learning Frameworks

被引:33
作者
Zhang, Lefeng [1 ,2 ]
Zhu, Tianqing [1 ,2 ]
Zhang, Haibin [3 ]
Xiong, Ping [4 ]
Zhou, Wanlei [1 ,2 ]
机构
[1] Univ Technol, Centrefor Cyber Secur & Privacy, Sydney, NSW 2007, Australia
[2] Univ Technol, Sch Comp Sci, Sydney, NSW 2007, Australia
[3] Beijing Univ Technol, Fac Sci, Dept Operat Res & Sci Computat, Beijing 100021, Peoples R China
[4] Zhongnan Univ Econ & Law, Sch Informat & Safety Engn, Dept Comp Sci & Technol, Wuhan 430073, Peoples R China
基金
澳大利亚研究理事会;
关键词
Index Terms- Machine unlearning; differential privacy; feder-ated learning;
D O I
10.1109/TIFS.2023.3297905
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Over the past decades, the abundance of personal data has led to the rapid development of machine learning models and important advances in artificial intelligence (AI). However, alongside all the achievements, there are increasing privacy threats and security risks that may cause significant losses for data providers. Recent legislation requires that the private information about a user should be removed from a database as well as machine learning models upon certain deletion requests. While erasing data records from memory storage is straightforward, it is often challenging to remove the influence of particular data samples from a model that has already been trained. Machine unlearning is an emerging paradigm that aims to make machine learning models "forget" what they have learned about particular data. Nevertheless, the unlearning issue for federated learning has not been completely addressed due to its special working mode. First, existing solutions crucially rely on retraining-based model calibration, which is likely unavailable and can pose new privacy risks for federated learning frameworks. Second, today's efficient unlearning strategies are mainly designed for convex problems, which are incapable of handling more complicated learning tasks like neural networks. To overcome these limitations, we took advantage of differential privacy and developed an efficient machine unlearning algorithm named FedRecovery. The FedRecovery erases the impact of a client by removing a weighted sum of gradient residuals from the global model, and tailors the Gaussian noise to make the unlearned model and retrained model statistically indistinguishable. Furthermore, the algorithm neither requires retraining-based fine-tuning nor needs the assumption of convexity. Theoretical analyses show the rigorous indistinguishability guarantee. Additionally, the experiment results on real-world datasets demonstrate that the FedRecovery is efficient and is able to produce a model that performs similarly to the retrained one.
引用
收藏
页码:4732 / 4746
页数:15
相关论文
共 51 条
[1]  
[Anonymous], 2016, Official Journal of the European Union
[2]  
Bhagoji AN, 2019, PR MACH LEARN RES, V97
[3]  
Bollapragada R, 2018, PR MACH LEARN RES, V80
[4]   Machine Unlearning [J].
Bourtoule, Lucas ;
Chandrasekaran, Varun ;
Choquette-Choo, Christopher A. ;
Jia, Hengrui ;
Travers, Adelin ;
Zhang, Baiwu ;
Lie, David ;
Papernot, Nicolas .
2021 IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP, 2021, :141-159
[5]   Concentrated Differential Privacy: Simplifications, Extensions, and Lower Bounds [J].
Bun, Mark ;
Steinke, Thomas .
THEORY OF CRYPTOGRAPHY, TCC 2016-B, PT I, 2016, 9985 :635-658
[6]   Towards Making Systems Forget with Machine Unlearning [J].
Cao, Yinzhi ;
Yang, Junfeng .
2015 IEEE SYMPOSIUM ON SECURITY AND PRIVACY SP 2015, 2015, :463-480
[7]  
Chaudhuri K, 2011, J MACH LEARN RES, V12, P1069
[8]   Recommendation Unlearning [J].
Chen, Chong ;
Sun, Fei ;
Zhang, Min ;
Ding, Bolin .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :2768-2777
[9]  
Chizat L, 2019, ADV NEUR IN, V32
[10]  
Dwork C, 2006, LECT NOTES COMPUT SC, V4052, P1