FedMUA: Exploring the Vulnerabilities of Federated Learning to Malicious Unlearning Attacks

被引:0
|
作者
Chen, Jian [1 ]
Lin, Zehui [1 ]
Lin, Wanyu [1 ,2 ]
Shi, Wenlong [3 ]
Yin, Xiaoyan [4 ]
Wang, Di [5 ]
机构
[1] Hong Kong Polytech Univ, Dept Data Sci & Artificial Intelligence, Hong Kong, Peoples R China
[2] Hong Kong Polytech Univ, Dept Comp, Hong Kong, Peoples R China
[3] Huazhong Univ Sci & Technol, Sch Elect Informat & Commun, Wuhan 430074, Peoples R China
[4] Northwest Univ, Sch Informat Sci & Technol, Xian 710069, Peoples R China
[5] King Abdullah Univ Sci & Technol, Div Comp Elect & Math Sci & Engn, Thuwal 23955, Saudi Arabia
关键词
Predictive models; Data models; Servers; Federated learning; Computational modeling; Training; Training data; Robustness; General Data Protection Regulation; Distributed databases; unlearning attacks; targeted attacks;
D O I
10.1109/TIFS.2025.3531141
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Recently, the practical needs of "the right to be forgotten" in federated learning gave birth to a paradigm known as federated unlearning, which enables the server to forget personal data upon the client's removal request. Existing studies on federated unlearning have primarily focused on efficiently eliminating the influence of requested data from the client's model without retraining from scratch, however, they have rarely doubted the reliability of the global model posed by the discrepancy between its prediction performance before and after unlearning. To bridge this gap, we take the first step by introducing a novel malicious unlearning attack dubbed FedMUA, aiming to unveil potential vulnerabilities emerging from federated learning during the unlearning process. Specifically, clients may act as attackers by crafting malicious unlearning requests to manipulate the prediction behavior of the global model. The crux of FedMUA is to mislead the global model into unlearning more information associated with the influential samples for the target sample than anticipated, thus inducing adverse effects on target samples from other clients. To achieve this, we design a novel two-step method, known as Influential Sample Identification and Malicious Unlearning Generation, to identify and subsequently generate malicious feature unlearning requests within the influential samples. By doing so, we can significantly alter the predictions pertaining to the target sample by initiating the malicious feature unlearning requests, leading to the deliberate manipulation for the user adversely. Additionally, we design a new defense mechanism that is highly resilient against malicious unlearning attacks. Extensive experiments on three realistic datasets reveal that FedMUA effectively induces misclassification on target samples and can achieve an 80% attack success rate by triggering only 0.3% malicious unlearning requests.
引用
收藏
页码:1665 / 1678
页数:14
相关论文
共 50 条
  • [21] Toward Efficient and Certified Recovery From Poisoning Attacks in Federated Learning
    Jiang, Yu
    Shen, Jiyuan
    Liu, Ziyao
    Tan, Chee Wei
    Lam, Kwok-Yan
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 : 2632 - 2647
  • [22] Label-Flipping Attacks in GNN-Based Federated Learning
    Yu, Shanqing
    Shen, Jie
    Xu, Shaocong
    Wang, Jinhuan
    Wang, Zeyu
    Xuan, Qi
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2025, 12 (02): : 1357 - 1368
  • [23] Secure Model Aggregation Against Poisoning Attacks for Cross-Silo Federated Learning With Robustness and Fairness
    Mao, Yunlong
    Ye, Zhujing
    Yuan, Xinyu
    Zhong, Sheng
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 6321 - 6336
  • [24] Compressed Particle-Based Federated Bayesian Learning and Unlearning
    Gong, Jinu
    Simeone, Osvaldo
    Kang, Joonhyuk
    IEEE COMMUNICATIONS LETTERS, 2023, 27 (02) : 556 - 560
  • [25] Adaptive federated learning scheme for recognition of malicious attacks in an IoT network
    Chhikara, Prateek
    Tekchandani, Rajkumar
    Kumar, Neeraj
    COMPUTING, 2024, 106 (04) : 1177 - 1192
  • [26] Adaptive federated learning scheme for recognition of malicious attacks in an IoT network
    Prateek Chhikara
    Rajkumar Tekchandani
    Neeraj Kumar
    Computing, 2024, 106 : 1177 - 1192
  • [27] Adaptive Selection of Loss Function for Federated Learning Clients Under Adversarial Attacks
    Lee, Suchul
    IEEE ACCESS, 2024, 12 : 96051 - 96062
  • [28] RobustFL: Robust Federated Learning Against Poisoning Attacks in Industrial IoT Systems
    Zhang, Jiale
    Ge, Chunpeng
    Hu, Feng
    Chen, Bing
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (09) : 6388 - 6397
  • [29] SPFL: A Self-Purified Federated Learning Method Against Poisoning Attacks
    Liu, Zizhen
    He, Weiyang
    Chang, Chip-Hong
    Ye, Jing
    Li, Huawei
    Li, Xiaowei
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 6604 - 6619
  • [30] Poisoning Attacks in Federated Learning: A Survey
    Xia, Geming
    Chen, Jian
    Yu, Chaodong
    Ma, Jun
    IEEE ACCESS, 2023, 11 : 10708 - 10722