Robustness Evaluations of Sustainable Machine Learning Models against Data Poisoning Attacks in the Internet of Things

被引:46
作者
Dunn, Corey [1 ]
Moustafa, Nour [1 ]
Turnbull, Benjamin [1 ]
机构
[1] Univ New South Wales, Sch Engn & Informat Technol, Canberra, ACT 2612, Australia
关键词
adversarial machine learning; sustainable machine learning; data poisoning; deep learning; Internet of Things; IOT; SECURITY;
D O I
10.3390/su12166434
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
With the increasing popularity of the Internet of Things (IoT) platforms, the cyber security of these platforms is a highly active area of research. One key technology underpinning smart IoT systems is machine learning, which classifies and predicts events from large-scale data in IoT networks. Machine learning is susceptible to cyber attacks, particularly data poisoning attacks that inject false data when training machine learning models. Data poisoning attacks degrade the performances of machine learning models. It is an ongoing research challenge to develop trustworthy machine learning models resilient and sustainable against data poisoning attacks in IoT networks. We studied the effects of data poisoning attacks on machine learning models, including the gradient boosting machine, random forest, naive Bayes, and feed-forward deep learning, to determine the levels to which the models should be trusted and said to be reliable in real-world IoT settings. In the training phase, a label modification function is developed to manipulate legitimate input classes. The function is employed at data poisoning rates of 5%, 10%, 20%, and 30% that allow the comparison of the poisoned models and display their performance degradations. The machine learning models have been evaluated using the ToN_IoT and UNSW NB-15 datasets, as they include a wide variety of recent legitimate and attack vectors. The experimental results revealed that the models' performances will be degraded, in terms of accuracy and detection rates, if the number of the trained normal observations is not significantly larger than the poisoned data. At the rate of data poisoning of 30% or greater on input data, machine learning performances are significantly degraded.
引用
收藏
页数:17
相关论文
共 39 条
[1]  
Ahlers D, 2019, COMM COM INF SC, V1041, P109, DOI 10.1007/978-3-030-22482-0_9
[2]   A Survey of Machine and Deep Learning Methods for Internet of Things (IoT) Security [J].
Al-Garadi, Mohammed Ali ;
Mohamed, Amr ;
Al-Ali, Abdulla Khalid ;
Du, Xiaojiang ;
Ali, Ihsan ;
Guizani, Mohsen .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2020, 22 (03) :1646-1685
[3]  
[Anonymous], 2014, P 2014 WORKSH ART IN, DOI DOI 10.1145/2666652.2666666
[4]  
[Anonymous], 2009, MACHINE LEARNING CYB, DOI DOI 10.1007/978-1-60327-555-2_2
[5]  
[Anonymous], 2017, ARXIV170301340
[6]   The security of machine learning [J].
Barreno, Marco ;
Nelson, Blaine ;
Joseph, Anthony D. ;
Tygar, J. D. .
MACHINE LEARNING, 2010, 81 (02) :121-148
[7]  
Biggio B., 2012, INT C MACHINE LEARNI
[8]   Enabling IoT Ecosystems through Platform Interoperability [J].
Broring, Arne ;
Schmid, Stefan ;
Schindhelm, Corina-Kim ;
Khelil, Abdelmajid ;
Kabisch, Sebastian ;
Kramer, Denis ;
Danh Le Phuoc ;
Mitic, Jelena ;
Anicic, Darko ;
Teniente, Ernest .
IEEE SOFTWARE, 2017, 34 (01) :54-61
[9]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[10]   A Survey of Adversarial Machine Learning in Cyber Warfare [J].
Duddu, Vasisht .
DEFENCE SCIENCE JOURNAL, 2018, 68 (04) :356-366