A Framework for Robust Deep Learning Models Against Adversarial Attacks Based on a Protection Layer Approach

被引:1
作者
Al-Andoli, Mohammed Nasser [1 ]
Tan, Shing Chiang [2 ]
Sim, Kok Swee [3 ]
Goh, Pey Yun [2 ]
Lim, Chee Peng [4 ]
机构
[1] Univ Teknikal Malaysia Melaka, Fac Informat & Commun Technol, Durian Tunggal 76100, Malaysia
[2] Multimedia Univ, Fac Informat Sci & Technol, Melaka 75450, Malaysia
[3] Multimedia Univ, Fac Engn & Technol, Melaka 75450, Malaysia
[4] Deakin Univ, Inst Intelligent Syst Res & Innovat, Waurn Ponds, Vic 3216, Australia
关键词
Deep learning; Adversarial machine learning; adversarial examples; security; adversarial attacks; adversarial examples detection; COMMUNITY DETECTION;
D O I
10.1109/ACCESS.2024.3354699
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning (DL) has demonstrated remarkable achievements in various fields. Nevertheless, DL models encounter significant challenges in detecting and defending against adversarial samples (AEs). These AEs are meticulously crafted by adversaries, introducing imperceptible perturbations to clean data to deceive DL models. Consequently, AEs pose potential risks to DL applications. In this paper, we propose an effective framework for enhancing the robustness of DL models against adversarial attacks. The framework leverages convolutional neural networks (CNNs) for feature learning, Deep Neural Networks (DNNs) with softmax for classification, and a defense mechanism to identify and exclude AEs. Evasion attacks are employed to create AEs to evade and mislead the classifier by generating malicious samples during the test phase of DL models i.e., CNN and DNN, using the Fast Gradient Sign Method (FGSM), Basic Iterative Method (BIM), Projected Gradient Descent (PGD), and Square Attack (SA). A protection layer is developed as a detection mechanism placed before the DNN classifier to identify and exclude AEs. The detection mechanism incorporates a machine learning model, which includes one of the following: Fuzzy ARTMAP, Random Forest, K-Nearest Neighbors, XGBoost, or Gradient Boosting Machine. Extensive evaluations are conducted on the MNIST, CIFAR-10, SVHN, and Fashion-MNIST data sets to assess the effectiveness of the proposed framework. The experimental results indicate the framework's ability to effectively and accurately detect AEs generated by four popular attacking methods, highlighting the potential of our developed framework in enhancing its robustness against AEs.
引用
收藏
页码:17522 / 17540
页数:19
相关论文
共 50 条
  • [21] Deep Learning Defense Method Against Adversarial Attacks
    Wang, Ling
    Zhang, Cheng
    Liu, Jie
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 3667 - 3671
  • [22] How to Defend and Secure Deep Learning Models Against Adversarial Attacks in Computer Vision: A Systematic Review
    Dhamija, Lovi
    Bansal, Urvashi
    [J]. NEW GENERATION COMPUTING, 2024, 42 (05) : 1165 - 1235
  • [23] Adversarial attacks on deep learning models in smart grids
    Hao, Jingbo
    Tao, Yang
    [J]. ENERGY REPORTS, 2022, 8 : 123 - 129
  • [24] Robust Deep Learning Models against Semantic-Preserving Adversarial Attack
    Zhao, Yunce
    Gao, Dashan
    Yao, Yinghua
    Zhang, Zeqi
    Mao, Bifei
    Yao, Xin
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [25] Mockingbird: Defending Against Deep-Learning-Based Website Fingerprinting Attacks With Adversarial Traces
    Rahman, Mohammad Saidur
    Imani, Mohsen
    Mathews, Nate
    Wright, Matthew
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 (16) : 1594 - 1609
  • [26] Detection of adversarial attacks against security systems based on deep learning model
    Jaber, Mohanad J.
    Jaber, Zahraa Jasim
    Obaid, Ahmed J.
    [J]. JOURNAL OF DISCRETE MATHEMATICAL SCIENCES & CRYPTOGRAPHY, 2024, 27 (05) : 1523 - 1538
  • [27] Invisible Adversarial Attacks on Deep Learning-Based Face Recognition Models
    Lin, Chih-Yang
    Chen, Feng-Jie
    Ng, Hui-Fuang
    Lin, Wei-Yang
    [J]. IEEE ACCESS, 2023, 11 : 51567 - 51577
  • [28] ADVERSARIAL ATTACKS ON RADAR TARGET RECOGNITION BASED ON DEEP LEARNING
    Zhou, Jie
    Peng, Bo
    Peng, Bowen
    [J]. 2022 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2022), 2022, : 2646 - 2649
  • [29] Robust Audio Watermarking Against Manipulation Attacks Based on Deep Learning
    Wen, Shuangbing
    Zhang, Qishan
    Hu, Tao
    Li, Jun
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 126 - 130
  • [30] Adversarial Deep Learning: A Survey on Adversarial Attacks and Defense Mechanisms on Image Classification
    Khamaiseh, Samer Y.
    Bagagem, Derek
    Al-Alaj, Abdullah
    Mancino, Mathew
    Alomari, Hakam W.
    [J]. IEEE ACCESS, 2022, 10 : 102266 - 102291