A Framework for Robust Deep Learning Models Against Adversarial Attacks Based on a Protection Layer Approach

被引:1
|
作者
Al-Andoli, Mohammed Nasser [1 ]
Tan, Shing Chiang [2 ]
Sim, Kok Swee [3 ]
Goh, Pey Yun [2 ]
Lim, Chee Peng [4 ]
机构
[1] Univ Teknikal Malaysia Melaka, Fac Informat & Commun Technol, Durian Tunggal 76100, Malaysia
[2] Multimedia Univ, Fac Informat Sci & Technol, Melaka 75450, Malaysia
[3] Multimedia Univ, Fac Engn & Technol, Melaka 75450, Malaysia
[4] Deakin Univ, Inst Intelligent Syst Res & Innovat, Waurn Ponds, Vic 3216, Australia
关键词
Deep learning; Adversarial machine learning; adversarial examples; security; adversarial attacks; adversarial examples detection; COMMUNITY DETECTION;
D O I
10.1109/ACCESS.2024.3354699
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning (DL) has demonstrated remarkable achievements in various fields. Nevertheless, DL models encounter significant challenges in detecting and defending against adversarial samples (AEs). These AEs are meticulously crafted by adversaries, introducing imperceptible perturbations to clean data to deceive DL models. Consequently, AEs pose potential risks to DL applications. In this paper, we propose an effective framework for enhancing the robustness of DL models against adversarial attacks. The framework leverages convolutional neural networks (CNNs) for feature learning, Deep Neural Networks (DNNs) with softmax for classification, and a defense mechanism to identify and exclude AEs. Evasion attacks are employed to create AEs to evade and mislead the classifier by generating malicious samples during the test phase of DL models i.e., CNN and DNN, using the Fast Gradient Sign Method (FGSM), Basic Iterative Method (BIM), Projected Gradient Descent (PGD), and Square Attack (SA). A protection layer is developed as a detection mechanism placed before the DNN classifier to identify and exclude AEs. The detection mechanism incorporates a machine learning model, which includes one of the following: Fuzzy ARTMAP, Random Forest, K-Nearest Neighbors, XGBoost, or Gradient Boosting Machine. Extensive evaluations are conducted on the MNIST, CIFAR-10, SVHN, and Fashion-MNIST data sets to assess the effectiveness of the proposed framework. The experimental results indicate the framework's ability to effectively and accurately detect AEs generated by four popular attacking methods, highlighting the potential of our developed framework in enhancing its robustness against AEs.
引用
收藏
页码:17522 / 17540
页数:19
相关论文
共 50 条
  • [1] Defending Deep Learning Models Against Adversarial Attacks
    Mani, Nag
    Moh, Melody
    Moh, Teng-Sheng
    INTERNATIONAL JOURNAL OF SOFTWARE SCIENCE AND COMPUTATIONAL INTELLIGENCE-IJSSCI, 2021, 13 (01): : 72 - 89
  • [2] Robust Adversarial Objects against Deep Learning Models
    Tsai, Tzungyu
    Yang, Kaichen
    Ho, Tsung-Yi
    Jin, Yier
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 954 - 962
  • [3] Beyond accuracy and precision: a robust deep learning framework to enhance the resilience of face mask detection models against adversarial attacks
    Sheikh, Burhan Ul Haque
    Zafar, Aasim
    EVOLVING SYSTEMS, 2024, 15 (01) : 1 - 24
  • [4] Beyond accuracy and precision: a robust deep learning framework to enhance the resilience of face mask detection models against adversarial attacks
    Burhan Ul Haque sheikh
    Aasim Zafar
    Evolving Systems, 2024, 15 : 1 - 24
  • [5] ACADIA: Efficient and Robust Adversarial Attacks Against Deep Reinforcement Learning
    Ali, Haider
    Al Ameedi, Mohannad
    Swami, Ananthram
    Ning, Rui
    Li, Jiang
    Wu, Hongyi
    Cho, Jin-Hee
    2022 IEEE CONFERENCE ON COMMUNICATIONS AND NETWORK SECURITY (CNS), 2022, : 1 - 9
  • [6] Robust Deep Object Tracking against Adversarial Attacks
    Jia, Shuai
    Ma, Chao
    Song, Yibing
    Yang, Xiaokang
    Yang, Ming-Hsuan
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025, 133 (03) : 1238 - 1257
  • [7] Defense Against Adversarial Attacks in Deep Learning
    Li, Yuancheng
    Wang, Yimeng
    APPLIED SCIENCES-BASEL, 2019, 9 (01):
  • [8] Adversarially Enhanced Learning (AEL): Robust lightweight deep learning approach for radiology image classification against adversarial attacks
    Singh, Anshu
    Singh, Maheshwari Prasad
    Singh, Amit Kumar
    IMAGE AND VISION COMPUTING, 2025, 154
  • [9] Adversarial Attacks and Defenses for Deep Learning Models
    Li M.
    Jiang P.
    Wang Q.
    Shen C.
    Li Q.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2021, 58 (05): : 909 - 926
  • [10] Copyright protection framework for federated learning models against collusion attacks
    Luo, Yuling
    Li, Yuanze
    Qin, Sheng
    Fu, Qiang
    Liu, Junxiu
    INFORMATION SCIENCES, 2024, 680