Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW

被引:2
|
作者
Villegas-Ch, William [1 ]
Jaramillo-Alcazar, Angel [1 ]
Lujan-Mora, Sergio [2 ]
机构
[1] Univ Las Amer, Escuela Ingn Cibersegur, Fac Ingn Ciencias Aplicadas, Quito 170125, Ecuador
[2] Univ Alicante, Dept Lenguajes & Sistemas Informat, Alicante 03690, Spain
关键词
adversary examples; robustness of models; countermeasures; NEURAL-NETWORKS;
D O I
10.3390/bdcc8010008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study evaluated the generation of adversarial examples and the subsequent robustness of an image classification model. The attacks were performed using the Fast Gradient Sign method, the Projected Gradient Descent method, and the Carlini and Wagner attack to perturb the original images and analyze their impact on the model's classification accuracy. Additionally, image manipulation techniques were investigated as defensive measures against adversarial attacks. The results highlighted the model's vulnerability to conflicting examples: the Fast Gradient Signed Method effectively altered the original classifications, while the Carlini and Wagner method proved less effective. Promising approaches such as noise reduction, image compression, and Gaussian blurring were presented as effective countermeasures. These findings underscore the importance of addressing the vulnerability of machine learning models and the need to develop robust defenses against adversarial examples. This article emphasizes the urgency of addressing the threat posed by harmful standards in machine learning models, highlighting the relevance of implementing effective countermeasures and image manipulation techniques to mitigate the effects of adversarial attacks. These efforts are crucial to safeguarding model integrity and trust in an environment marked by constantly evolving hostile threats. An average 25% decrease in accuracy was observed for the VGG16 model when exposed to the Fast Gradient Signed Method and Projected Gradient Descent attacks, and an even more significant 35% decrease with the Carlini and Wagner method.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] Defense Against Adversarial Attacks in Deep Learning
    Li, Yuancheng
    Wang, Yimeng
    APPLIED SCIENCES-BASEL, 2019, 9 (01):
  • [2] Analyzing the Robustness of Deep Learning Against Adversarial Examples
    Zhao, Jun
    2018 56TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2018, : 1060 - 1064
  • [3] Robustness Against Adversarial Attacks Using Dimensionality
    Chattopadhyay, Nandish
    Chatterjee, Subhrojyoti
    Chattopadhyay, Anupam
    SECURITY, PRIVACY, AND APPLIED CRYPTOGRAPHY ENGINEERING, SPACE 2021, 2022, 13162 : 226 - 241
  • [4] Robustness of Deep Learning-Based Specific Emitter Identification under Adversarial Attacks
    Sun, Liting
    Ke, Da
    Wang, Xiang
    Huang, Zhitao
    Huang, Kaizhu
    REMOTE SENSING, 2022, 14 (19)
  • [5] How to Defend and Secure Deep Learning Models Against Adversarial Attacks in Computer Vision: A Systematic Review
    Dhamija, Lovi
    Bansal, Urvashi
    NEW GENERATION COMPUTING, 2024, 42 (05) : 1165 - 1235
  • [6] Physical Adversarial Attacks Against Deep Learning Based Channel Decoding Systems
    Babu, Surabhi Ashok
    Ameer, P. M.
    2020 IEEE REGION 10 SYMPOSIUM (TENSYMP) - TECHNOLOGY FOR IMPACTFUL SUSTAINABLE DEVELOPMENT, 2020, : 1511 - 1514
  • [7] Deep learning for steganalysis: evaluating model robustness against image transformations
    Alrusaini, Othman A.
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2025, 8
  • [8] A Methodology for Evaluating the Robustness of Anomaly Detectors to Adversarial Attacks in Industrial Scenarios
    Perales Gomez, Angel Luis
    Fernandez Maimo, Lorenzo
    Garcia Clemente, Felix J.
    Maroto Morales, Javier Alejandro
    Huertas Celdran, Alberto
    Bovet, Gerome
    IEEE ACCESS, 2022, 10 : 124582 - 124594
  • [9] A survey on robustness attacks for deep code models
    Qu, Yubin
    Huang, Song
    Yao, Yongming
    AUTOMATED SOFTWARE ENGINEERING, 2024, 31 (02)
  • [10] Invisible Adversarial Attacks on Deep Learning-Based Face Recognition Models
    Lin, Chih-Yang
    Chen, Feng-Jie
    Ng, Hui-Fuang
    Lin, Wei-Yang
    IEEE ACCESS, 2023, 11 : 51567 - 51577