Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW

被引:2
|
作者
Villegas-Ch, William [1 ]
Jaramillo-Alcazar, Angel [1 ]
Lujan-Mora, Sergio [2 ]
机构
[1] Univ Las Amer, Escuela Ingn Cibersegur, Fac Ingn Ciencias Aplicadas, Quito 170125, Ecuador
[2] Univ Alicante, Dept Lenguajes & Sistemas Informat, Alicante 03690, Spain
关键词
adversary examples; robustness of models; countermeasures; NEURAL-NETWORKS;
D O I
10.3390/bdcc8010008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study evaluated the generation of adversarial examples and the subsequent robustness of an image classification model. The attacks were performed using the Fast Gradient Sign method, the Projected Gradient Descent method, and the Carlini and Wagner attack to perturb the original images and analyze their impact on the model's classification accuracy. Additionally, image manipulation techniques were investigated as defensive measures against adversarial attacks. The results highlighted the model's vulnerability to conflicting examples: the Fast Gradient Signed Method effectively altered the original classifications, while the Carlini and Wagner method proved less effective. Promising approaches such as noise reduction, image compression, and Gaussian blurring were presented as effective countermeasures. These findings underscore the importance of addressing the vulnerability of machine learning models and the need to develop robust defenses against adversarial examples. This article emphasizes the urgency of addressing the threat posed by harmful standards in machine learning models, highlighting the relevance of implementing effective countermeasures and image manipulation techniques to mitigate the effects of adversarial attacks. These efforts are crucial to safeguarding model integrity and trust in an environment marked by constantly evolving hostile threats. An average 25% decrease in accuracy was observed for the VGG16 model when exposed to the Fast Gradient Signed Method and Projected Gradient Descent attacks, and an even more significant 35% decrease with the Carlini and Wagner method.
引用
收藏
页数:23
相关论文
共 50 条
  • [31] Adversarial Attacks and Countermeasures on Image Classification-based Deep Learning Models in Autonomous Driving Systems: A Systematic Review
    Badjie, Bakary
    Cecilio, Jose
    Casimiro, Antonio
    ACM COMPUTING SURVEYS, 2025, 57 (01)
  • [32] Adversarial Attacks on Deep Learning Based Power Allocation in a Massive MIMO Network
    Manoj, B. R.
    Sadeghi, Meysam
    Larsson, Erik G.
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [33] Adversarial Deception in Deep Learning: Analysis and Mitigation
    Wei, Wenqi
    Liu, Ling
    Loper, Margaret
    Chow, Ka-Ho
    Gursoy, Mehmet Emre
    Truex, Stacey
    Wu, Yanzhao
    2020 SECOND IEEE INTERNATIONAL CONFERENCE ON TRUST, PRIVACY AND SECURITY IN INTELLIGENT SYSTEMS AND APPLICATIONS (TPS-ISA 2020), 2020, : 236 - 245
  • [34] Protection against Adversarial Attacks on Malware Detectors Using Machine Learning Algorithms
    Marshev, I. I.
    Zhukovskii, E., V
    Aleksandrova, E. B.
    AUTOMATIC CONTROL AND COMPUTER SCIENCES, 2021, 55 (08) : 1025 - 1028
  • [35] Approach to Detecting Attacks against Machine Learning Systems with a Generative Adversarial Network
    Kotenko, I. V.
    Saenko, I. B.
    Lauta, O. S.
    Vasilev, N. A.
    Sadovnikov, V. E.
    PATTERN RECOGNITION AND IMAGE ANALYSIS, 2024, 34 (03) : 589 - 596
  • [36] Protection against Adversarial Attacks on Malware Detectors Using Machine Learning Algorithms
    I. I. Marshev
    E. V. Zhukovskii
    E. B. Aleksandrova
    Automatic Control and Computer Sciences, 2021, 55 : 1025 - 1028
  • [37] Security Hardening of Intelligent Reflecting Surfaces Against Adversarial Machine Learning Attacks
    Catak, Ferhat Ozgur
    Kuzlu, Murat
    Tang, Haolin
    Catak, Evren
    Zhao, Yanxiao
    IEEE ACCESS, 2022, 10 : 100267 - 100275
  • [38] Evaluating deep learning models for sentiment classification
    Karakus, Betul Ay
    Talo, Muhammed
    Hallac, Ibrahim Riza
    Aydin, Galip
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2018, 30 (21)
  • [39] Generative Adversarial Attacks Against Intrusion Detection Systems Using Active Learning
    Shu, Dule
    Leslie, Nandi O.
    Kamhoua, Charles A.
    Tucker, Conrad S.
    PROCEEDINGS OF THE 2ND ACM WORKSHOP ON WIRELESS SECURITY AND MACHINE LEARNING, WISEML 2020, 2020, : 1 - 6
  • [40] TnT Attacks! Universal Naturalistic Adversarial Patches Against Deep Neural Network Systems
    Doan, Bao Gia
    Xue, Minhui
    Ma, Shiqing
    Abbasnejad, Ehsan
    Ranasinghe, Damith C.
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2022, 17 : 3816 - 3830