Adversarial Attacks on Featureless Deep Learning Malicious URLs Detection

被引:9
作者
Rasheed, Bader [1 ]
Khan, Adil [1 ]
Kazmi, S. M. Ahsan [2 ]
Hussain, Rasheed [2 ]
Piran, Md Jalil [3 ]
Suh, Doug Young [4 ]
机构
[1] Innopolis Univ, Inst Data Sci & Artificial Intelligence, Innopolis 420500, Russia
[2] Innopolis Univ, Inst Informat Secur & Cyberphys Syst, Innopolis 420500, Russia
[3] Sejong Univ, Dept Comp Sci & Engn, Seoul, South Korea
[4] Kyung Hee Univ, Dept Elect Engn, Yongin, South Korea
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2021年 / 68卷 / 01期
关键词
Malicious URLs; detection; deep learning; adversarial attack; web security; DEFENSE;
D O I
10.32604/cmc.2021.015452
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Detecting malicious Uniform Resource Locators (URLs) is crucially important to prevent attackers from committing cybercrimes. Recent researches have investigated the role of machine learning (ML) models to detect malicious URLs. By using ML algorithms, first, the features of URLs are extracted, and then different ML models are trained. The limitation of this approach is that it requires manual feature engineering and it does not consider the sequential patterns in the URL. Therefore, deep learning (DL) models are used to solve these issues since they are able to perform featureless detection. Furthermore, DL models give better accuracy and generalization to newly designed URLs; however, the results of our study show that these models, such as any other DL models, can be susceptible to adversarial attacks. In this paper, we examine the robustness of these models and demonstrate the importance of considering this susceptibility before applying such detection systems in real-world solutions. We propose and demonstrate a black-box attack based on scoring functions with greedy search for the minimum number of perturbations leading to a misclassification. The attack is examined against different types of convolutional neural networks (CNN)-based URL classifiers and it causes a tangible decrease in the accuracy with more than 56% reduction in the accuracy of the best classifier (among the selected classifiers for this work). Moreover, adversarial training shows promising results in reducing the influence of the attack on the robustness of the model to less than 7% on average.
引用
收藏
页码:921 / 939
页数:19
相关论文
共 30 条
[1]   Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey [J].
Akhtar, Naveed ;
Mian, Ajmal .
IEEE ACCESS, 2018, 6 :14410-14430
[2]  
Akinsanya B. J., 2019, SOFTWARE TESTING MAC, V1, P1
[3]   Bypassing Detection of URL-based Phishing Attacks Using Generative Adversarial Deep Neural Networks [J].
AlEroud, Ahmed ;
Karabatis, George .
PROCEEDINGS OF THE SIXTH INTERNATIONAL WORKSHOP ON SECURITY AND PRIVACY ANALYTICS (IWSPA'20), 2020, :53-60
[4]  
[Anonymous], 2018, Adversarial attacks and defences: A survey
[5]  
[Anonymous], 2014, PROC C EMPIRICAL MET, DOI DOI 10.3115/V1/D14-1181
[6]  
[Anonymous], 2019, MALICIOUS BENIGN WEB
[7]  
AWPG, 2020, PHISH ACT TRENDS REP
[8]   Accident Recognition via 3D CNNs for Automated Traffic Monitoring in Smart Cities [J].
Bortnikov, Mikhail ;
Khan, Adil ;
Khattak, Asad Masood ;
Ahmad, Muhammad .
ADVANCES IN COMPUTER VISION, VOL 2, 2020, 944 :256-264
[9]   Using Adversarial Examples to Bypass Deep Learning Based URL Detection System [J].
Chen, Wencheng ;
Zeng, Yi ;
Qiu, Meikang .
4TH IEEE INTERNATIONAL CONFERENCE ON SMART CLOUD (SMARTCLOUD 2019) / 3RD INTERNATIONAL SYMPOSIUM ON REINFORCEMENT LEARNING (ISRL 2019), 2019, :128-130
[10]  
Dos Santos C., 2014, Proceedings of COLING 2014, the 25th international conference on computational linguistics: technical papers, P69