Investigating the Factors Impacting Adversarial Attack and Defense Performances in Federated Learning

被引:4
作者
Aljaafari, Nura [1 ]
Nazzal, Mahmoud [2 ]
Sawalmeh, Ahmad H.
Khreishah, Abdallah [2 ]
Anan, Muhammad [3 ]
Algosaibi, Abdulelah [1 ]
Alnaeem, Mohammed Abdulaziz [4 ]
Aldalbahi, Adel [1 ]
Alhumam, Abdulaziz [1 ]
Vizcarra, Conrado P. [1 ]
机构
[1] King Faisal Univ, Dept Comp Sci, Al Hufuf 31982, Saudi Arabia
[2] New Jersey Inst Technol, Dept Elect & Comp Engn, Newark, NJ 07102 USA
[3] Alfaisal Univ, Dept Software Engn, Riyadh 11533, Saudi Arabia
[4] King Faisal Univ, Dept Comp Networks & Commun, Al Hufuf 31982, Saudi Arabia
关键词
Training; Data models; Task analysis; Analytical models; Servers; Computational modeling; Complexity theory; Adversarial attacks; adversarial defense; federated learning; machine learning security; PRIVACY;
D O I
10.1109/TEM.2022.3155353
中图分类号
F [经济];
学科分类号
02 ;
摘要
Despite the promising success of federated learning in various application areas, its inherent vulnerability to adversarial attacks hinders its applicability in security-critical areas. This calls for developing viable defense measures against such attacks. A prerequisite for this development, however, is the understanding of what creates, promotes, and aggravates this vulnerability. To date, developing this understanding remains an outstanding gap in the literature. Accordingly, this paper presents an attempt at developing such an understanding. Primarily, this is achieved from two main perspectives. The first perspective concerns addressing the factors, elements, and parameters contributing to the vulnerability of federated learning models to adversarial attacks, their degrees of severity, and combined effects. This includes addressing diverse operating conditions, attack types and scenarios, and collaborations between attacking agents. The second perspective regards analyzing the appearance of the adversarial property of a model in how it updates its coefficients and exploiting this for defense purposes. These analyses are conducted through extensive experiments on image and text classification tasks. Simulation results reveal the importance of specific parameters and factors on the severity of this vulnerability. Besides, the proposed defense strategy is shown able to provide promising performances.
引用
收藏
页码:12542 / 12555
页数:14
相关论文
共 50 条
  • [21] Evil vs evil: using adversarial examples to against backdoor attack in federated learning
    Liu, Tao
    Li, Mingjun
    Zheng, Haibin
    Ming, Zhaoyan
    Chen, Jinyin
    MULTIMEDIA SYSTEMS, 2023, 29 (02) : 553 - 568
  • [22] Adversarial Node Placement in Decentralized Federated Learning: Maximum Spanning-Centrality Strategy and Performance Analysis
    Piaseczny, Adam
    Ruzomberka, Eric
    Parasnis, Rohit
    Brinton, Christopher G.
    IEEE INTERNET OF THINGS JOURNAL, 2025, 12 (01): : 45 - 60
  • [23] Exploring Adversarial Attacks in Federated Learning for Medical Imaging
    Darzi, Erfan
    Dubost, Florian
    Sijtsema, Nanna. M.
    van Ooijen, P. M. A.
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (12) : 13591 - 13599
  • [24] PerFED-GAN: Personalized Federated Learning via Generative Adversarial Networks
    Cao, Xingjian
    Sun, Gang
    Yu, Hongfang
    Guizani, Mohsen
    IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (05): : 3749 - 3762
  • [25] Swarm Optimization-Based Federated Learning for the Cyber Resilience of Internet of Things Systems Against Adversarial Attacks
    Yamany, Waleed
    Keshk, Marwa
    Moustafa, Nour
    Turnbull, Benjamin
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2024, 70 (01) : 1359 - 1369
  • [26] Backdoor attack and defense in federated generative adversarial network-based medical image synthesis
    Jin, Ruinan
    Li, Xiaoxiao
    MEDICAL IMAGE ANALYSIS, 2023, 90
  • [27] CapsuleBD: A Backdoor Attack Method Against Federated Learning Under Heterogeneous Models
    Liao, Yuying
    Zhao, Xuechen
    Zhou, Bin
    Huang, Yanyi
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 : 4071 - 4086
  • [28] Adversarial Attack and Defense on Deep Learning for Air Transportation Communication Jamming
    Liu, Mingqian
    Zhang, Zhenju
    Chen, Yunfei
    Ge, Jianhua
    Zhao, Nan
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (01) : 973 - 986
  • [29] LFGurad: A Defense against Label Flipping Attack in Federated Learning for Vehicular Network
    Sameera, K. M.
    Vinod, P.
    Rehiman, K. A. Rafidha
    Conti, Mauro
    COMPUTER NETWORKS, 2024, 254
  • [30] Privacy for Free: Spy Attack in Vertical Federated Learning by Both Active and Passive Parties
    Fu, Chaohao
    Chen, Hongbin
    Ruan, Na
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 : 2550 - 2563