Explainability and Adversarial Robustness for RNNs

被引:16
作者
Hartl, Alexander [1 ]
Bachl, Maximilian [1 ]
Fabini, Joachim [1 ]
Zseby, Tanja [1 ]
机构
[1] Tech Univ Wien, Vienna, Austria
来源
2020 IEEE SIXTH INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING SERVICE AND APPLICATIONS (BIGDATASERVICE 2020) | 2020年
关键词
NEURAL-NETWORKS;
D O I
10.1109/BigDataService49289.2020.00030
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recurrent Neural Networks (RNNs) yield attractive properties for constructing Intrusion Detection Systems (IDSs) for network data. With the rise of ubiquitous Machine Learning (ML) systems, malicious actors have been catching up quickly to find new ways to exploit ML vulnerabilities for profit. Recently developed adversarial ML techniques focus on computer vision and their applicability to network traffic is not straightforward: Network packets expose fewer features than an image, are sequential and impose several constraints on their features. We show that despite these completely different characteristics, adversarial samples can be generated reliably for RNNs. To understand a classifier's potential for misclassification, we extend existing explainability techniques and propose new ones, suitable particularly for sequential data. Applying them shows that already the first packets of a communication flow are of crucial importance and are likely to be targeted by attackers. Feature importance methods show that even relatively unimportant features can be effectively abused to generate adversarial samples. We thus introduce the concept of feature sensitivity which quantifies how much potential a feature has to cause misclassification. Since traditional evaluation metrics such as accuracy are not sufficient for quantifying the adversarial threat, we propose the Adversarial Robustness Score (ARS) for comparing IDSs and show that an adversarial training procedure can significantly and successfully reduce the attack surface.
引用
收藏
页码:149 / 157
页数:9
相关论文
共 24 条
  • [1] Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey
    Akhtar, Naveed
    Mian, Ajmal
    [J]. IEEE ACCESS, 2018, 6 : 14410 - 14430
  • [2] [Anonymous], 2017, ARXIV171208645
  • [3] [Anonymous], 2018, ICLR
  • [4] Walling up Backdoors in Intrusion Detection Systems
    Bachl, Maximilian
    Hartl, Alexander
    Fabini, Joachim
    Zseby, Tanja
    [J]. BIG-DAMA'19: PROCEEDINGS OF THE 3RD ACM CONEXT WORKSHOP ON BIG DATA, MACHINE LEARNING AND ARTIFICIAL INTELLIGENCE FOR DATA COMMUNICATION NETWORKS, 2019, : 8 - 13
  • [5] Carlini N., 2019, ARXIV190206705
  • [6] Towards Evaluating the Robustness of Neural Networks
    Carlini, Nicholas
    Wagner, David
    [J]. 2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, : 39 - 57
  • [7] Dhurandhar A., 2018, MODEL AGNOSTIC CONTR
  • [8] Greedy function approximation: A gradient boosting machine
    Friedman, JH
    [J]. ANNALS OF STATISTICS, 2001, 29 (05) : 1189 - 1232
  • [9] Goodfellow Ian J., 2015, 3 INT C LEARN REPR I
  • [10] Towards Evaluation of NIDSs in Adversarial Setting
    Hashemi, Mohammad J.
    Cusack, Greg
    Keller, Eric
    [J]. BIG-DAMA'19: PROCEEDINGS OF THE 3RD ACM CONEXT WORKSHOP ON BIG DATA, MACHINE LEARNING AND ARTIFICIAL INTELLIGENCE FOR DATA COMMUNICATION NETWORKS, 2019, : 14 - 21