Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW

被引:2
|
作者
Villegas-Ch, William [1 ]
Jaramillo-Alcazar, Angel [1 ]
Lujan-Mora, Sergio [2 ]
机构
[1] Univ Las Amer, Escuela Ingn Cibersegur, Fac Ingn Ciencias Aplicadas, Quito 170125, Ecuador
[2] Univ Alicante, Dept Lenguajes & Sistemas Informat, Alicante 03690, Spain
关键词
adversary examples; robustness of models; countermeasures; NEURAL-NETWORKS;
D O I
10.3390/bdcc8010008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study evaluated the generation of adversarial examples and the subsequent robustness of an image classification model. The attacks were performed using the Fast Gradient Sign method, the Projected Gradient Descent method, and the Carlini and Wagner attack to perturb the original images and analyze their impact on the model's classification accuracy. Additionally, image manipulation techniques were investigated as defensive measures against adversarial attacks. The results highlighted the model's vulnerability to conflicting examples: the Fast Gradient Signed Method effectively altered the original classifications, while the Carlini and Wagner method proved less effective. Promising approaches such as noise reduction, image compression, and Gaussian blurring were presented as effective countermeasures. These findings underscore the importance of addressing the vulnerability of machine learning models and the need to develop robust defenses against adversarial examples. This article emphasizes the urgency of addressing the threat posed by harmful standards in machine learning models, highlighting the relevance of implementing effective countermeasures and image manipulation techniques to mitigate the effects of adversarial attacks. These efforts are crucial to safeguarding model integrity and trust in an environment marked by constantly evolving hostile threats. An average 25% decrease in accuracy was observed for the VGG16 model when exposed to the Fast Gradient Signed Method and Projected Gradient Descent attacks, and an even more significant 35% decrease with the Carlini and Wagner method.
引用
收藏
页数:23
相关论文
共 50 条
  • [41] SIT: Stochastic Input Transformation to Defend Against Adversarial Attacks on Deep Neural Networks
    Guesmi, Amira
    Alouani, Ihsen
    Baklouti, Mouna
    Frikha, Tarek
    Abid, Mohamed
    IEEE DESIGN & TEST, 2022, 39 (03) : 63 - 72
  • [42] Detecting Adversarial Samples for Deep Learning Models: A Comparative Study
    Zhang, Shigeng
    Chen, Shuxin
    Liu, Xuan
    Hua, Chengyao
    Wang, Weiping
    Chen, Kai
    Zhang, Jian
    Wang, Jianxin
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2022, 9 (01): : 231 - 244
  • [43] ImageNet-Patch: A dataset for benchmarking machine learning robustness against adversarial patches
    Pintor, Maura
    Angioni, Daniele
    Sotgiu, Angelo
    Demetrio, Luca
    Demontis, Ambra
    Biggio, Battista
    Roli, Fabio
    PATTERN RECOGNITION, 2023, 134
  • [44] Tutorial: Toward Robust Deep Learning against Poisoning Attacks
    Chen, Huili
    Koushanfar, Farinaz
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2023, 22 (03)
  • [45] Jujutsu: A Two-stage Defense against Adversarial Patch Attacks on Deep Neural Networks
    Chen, Zitao
    Dash, Pritam
    Pattabiraman, Karthik
    PROCEEDINGS OF THE 2023 ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, ASIA CCS 2023, 2023, : 689 - 703
  • [46] Metamorphic Detection of Adversarial Examples in Deep Learning Models With Affine Transformations
    Mekala, Rohan Reddy
    Magnusson, Gudjon Einar
    Porter, Adam
    Lindvall, Mikael
    Diep, Madeline
    2019 IEEE/ACM 4TH INTERNATIONAL WORKSHOP ON METAMORPHIC TESTING (MET 2019), 2019, : 55 - 62
  • [47] Improving the accuracy and robustness of RRAM-based in-memory computing against RRAM hardware noise and adversarial attacks
    Cherupally, Sai Kiran
    Meng, Jian
    Rakin, Adnan Siraj
    Yin, Shihui
    Yeo, Injune
    Yu, Shimeng
    Fan, Deliang
    Seo, Jae-Sun
    SEMICONDUCTOR SCIENCE AND TECHNOLOGY, 2022, 37 (03)
  • [48] Adversarial attacks on deep learning networks in image classification based on Smell Bees Optimization Algorithm
    Ben Ali, Yamina Mohamed
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 140 : 185 - 195
  • [49] Server-Based Manipulation Attacks Against Machine Learning Models
    Liao, Cong
    Zhong, Haoti
    Zhu, Sencun
    Squicciarini, Anna
    PROCEEDINGS OF THE EIGHTH ACM CONFERENCE ON DATA AND APPLICATION SECURITY AND PRIVACY (CODASPY'18), 2018, : 24 - 34
  • [50] The Robustness of Modern Deep Learning Architectures against Single Event Upset Errors
    Arechiga, Austin P.
    Michaels, Alan J.
    2018 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2018,