An XAI-based adversarial training approach for cyber-threat detection

被引:3
作者
Al-Essa, Malik [1 ]
Andresini, Giuseppina [1 ]
Appice, Annalisa [1 ]
Malerba, Donato [1 ]
机构
[1] Univ Bari Aldo Moro, Dept Comp Sci, Bari, Italy
来源
2022 IEEE INTL CONF ON DEPENDABLE, AUTONOMIC AND SECURE COMPUTING, INTL CONF ON PERVASIVE INTELLIGENCE AND COMPUTING, INTL CONF ON CLOUD AND BIG DATA COMPUTING, INTL CONF ON CYBER SCIENCE AND TECHNOLOGY CONGRESS (DASC/PICOM/CBDCOM/CYBERSCITECH) | 2022年
关键词
Adversarial training; eXplainable Artificial Intelligence; Transfer learning; Cyber-threat detection;
D O I
10.1109/DASC/PiCom/CBDCom/Cy55231.2022.9927842
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Adversarial training is commonly used in the artificial intelligence literature to improve the robustness of deep neural models to adversarial samples. In addition, eXplainable Artificial Intelligence (XAI) has been recently investigated to improve the interpretability and explainability of black-box artificial systems such as deep neural models. In this study, we propose a methodology that combines adversarial training and XAI, in order to increase the accuracy of deep neural models trained for cyber-threat detection. In particular, we use the FGSM technique to generate the adversarial samples for the adversarial training stage, and SHAP to produce the local explanations of decisions made during the adversarial training stage. These local explanations are, subsequently, used to produce a new feature set that describes the effect of the original cyber-data characteristics on the classifications of the examples processed during the adversarial training stage. Leveraging this XAI-based information, we apply a transfer learning strategy, namely fine-tuning, to improve the accuracy performance of the deep neural model. Experiments conducted on two benchmark cybersecurity datasets prove the effectiveness of the proposed methodology in the multi-class classification of cyber-data.
引用
收藏
页码:806 / 813
页数:8
相关论文
共 24 条
  • [11] Goodfellow I.J., 2015, CoRR
  • [12] Khamis R. A., 2020, 2020 INT S NETW COMP, P1, DOI DOI 10.1109/ISNCC49221.2020.9297344
  • [13] Adversarial XAI Methods in Cybersecurity
    Kuppa, Aditya
    Nhien-An Le-Khac
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 : 4924 - 4938
  • [14] Classification and Explanation for Intrusion Detection System Based on Ensemble Trees and SHAP Method
    Le, Thi-Thu-Huong
    Kim, Haeyoung
    Kang, Hyoeun
    Kim, Howon
    [J]. SENSORS, 2022, 22 (03)
  • [15] Lundberg SM, 2017, ADV NEUR IN, V30
  • [16] Effective and Efficient Hybrid Android Malware Classification Using Pseudo-Label Stacked Auto-Encoder
    Mahdavifar, Samaneh
    Alhadidi, Dima
    Ghorbani, Ali. A.
    [J]. JOURNAL OF NETWORK AND SYSTEMS MANAGEMENT, 2022, 30 (01)
  • [17] Samek W., 2019, Explainable AI: Interpreting, Explaining and Visualizing Deep Learning
  • [18] Sarhan M., 2021, arXiv
  • [19] Achieving Explainability of Intrusion Detection System by Hybrid Oracle-Explainer Approach
    Szczepanski, Mateusz
    Choras, Michal
    Pawlicki, Marek
    Kozik, Rafal
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [20] Velickovic P., 2018, 6 INT C LEARN REPR I