Poisoning attacks on machine learning models in cyber systems and mitigation strategies

被引:1
|
作者
Izmailov, Rauf [1 ]
Venkatesan, Sridhar [1 ]
Reddy, Achyut [1 ]
Chadha, Ritu [1 ]
De Lucia, Michael [2 ]
Oprea, Alina [3 ]
机构
[1] Peraton Labs Inc, Basking Ridge, NJ 07920 USA
[2] DEVCOM Army Res Lab, Aberdeen Proving Ground, MD USA
[3] Northeastern Univ, Boston, MA 02115 USA
来源
DISRUPTIVE TECHNOLOGIES IN INFORMATION SCIENCES VI | 2022年 / 12117卷
关键词
Machine learning; adversarial machine learning; network intrusion detection; data poisoning; data cleaning; classifier; poisoning attack;
D O I
10.1117/12.2622112
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Poisoning attacks on training data are becoming one of the top concerns among users of machine learning systems. The goal of such attacks is to inject a small set of maliciously mislabeled training data into the training pipeline so as to detrimentally impact a machine learning model trained on such data. Constructing such attacks for cyber applications is especially challenging due to their realizability constraints. Furthermore, poisoning mitigation techniques for such applications are also not well understood. This paper investigates techniques for realizable data poisoning availability attacks (using several cyber applications), in which an attacker can insert a set of poisoned samples at the training time with the goal of degrading the accuracy of the deployed model. We design a white-box, realizable poisoning attack that degraded the original model's accuracy by generating mislabeled samples in close vicinity of a selected subset of training points. We investigate this strategy and its modifications for key classifier architectures and provide specific implications for each of them. The paper also proposes a novel data cleaning method as a defense against such poisoning attacks. Our defense includes a diversified ensemble of classifiers, each trained on a different subset of the training set. We use the disagreement of the classifiers' predictions as a decision whether to keep a given sample in the training dataset or remove it. The results demonstrate the efficiency of this strategy with very limited performance penalty.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Poisoning Attacks and Data Sanitization Mitigations for Machine Learning Models in Network Intrusion Detection Systems
    Venkatesan, Sridhar
    Sikka, Harshvardhan
    Izmailov, Rauf
    Chadha, Ritu
    Oprea, Alina
    de Lucia, Michael J.
    2021 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2021), 2021,
  • [2] GAN-Driven Data Poisoning Attacks and Their Mitigation in Federated Learning Systems
    Psychogyios, Konstantinos
    Velivassaki, Terpsichori-Helen
    Bourou, Stavroula
    Voulkidis, Artemis
    Skias, Dimitrios
    Zahariadis, Theodore
    ELECTRONICS, 2023, 12 (08)
  • [3] Data Poisoning Attacks on Federated Machine Learning
    Sun, Gan
    Cong, Yang
    Dong, Jiahua
    Wang, Qiang
    Lyu, Lingjuan
    Liu, Ji
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (13) : 11365 - 11375
  • [4] Model poisoning attacks against distributed machine learning systems
    Tomsett, Richard
    Chan, Kevin
    Chakraborty, Supriyo
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS, 2019, 11006
  • [5] The robustness of popular multiclass machine learning models against poisoning attacks: Lessons and insights
    Maabreh, Majdi
    Maabreh, Arwa
    Qolomany, Basheer
    Al-Fuqaha, Ala
    INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2022, 18 (07)
  • [6] Poisoning Attacks on Fair Machine Learning
    Minh-Hao Van
    Du, Wei
    Wu, Xintao
    Lu, Aidong
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2022, PT I, 2022, : 370 - 386
  • [7] Cyber Attacks Detection using Machine Learning in Smart Grid Systems
    Gyawali, Sohan
    Beg, Omar
    IEEE INFOCOM 2022 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2022,
  • [8] Data poisoning attacks against machine learning algorithms
    Yerlikaya, Fahri Anil
    Bahtiyar, Serif
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 208
  • [9] Threats to Training: A Survey of Poisoning Attacks and Defenses on Machine Learning Systems
    Wang, Zhibo
    Ma, Jingjing
    Wang, Xue
    Hu, Jiahui
    Qin, Zhan
    Ren, Kui
    ACM COMPUTING SURVEYS, 2023, 55 (07)
  • [10] Robustness Evaluations of Sustainable Machine Learning Models against Data Poisoning Attacks in the Internet of Things
    Dunn, Corey
    Moustafa, Nour
    Turnbull, Benjamin
    SUSTAINABILITY, 2020, 12 (16)