Adversarial Attacks on Featureless Deep Learning Malicious URLs Detection

被引:9
作者
Rasheed, Bader [1 ]
Khan, Adil [1 ]
Kazmi, S. M. Ahsan [2 ]
Hussain, Rasheed [2 ]
Piran, Md Jalil [3 ]
Suh, Doug Young [4 ]
机构
[1] Innopolis Univ, Inst Data Sci & Artificial Intelligence, Innopolis 420500, Russia
[2] Innopolis Univ, Inst Informat Secur & Cyberphys Syst, Innopolis 420500, Russia
[3] Sejong Univ, Dept Comp Sci & Engn, Seoul, South Korea
[4] Kyung Hee Univ, Dept Elect Engn, Yongin, South Korea
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2021年 / 68卷 / 01期
关键词
Malicious URLs; detection; deep learning; adversarial attack; web security; DEFENSE;
D O I
10.32604/cmc.2021.015452
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Detecting malicious Uniform Resource Locators (URLs) is crucially important to prevent attackers from committing cybercrimes. Recent researches have investigated the role of machine learning (ML) models to detect malicious URLs. By using ML algorithms, first, the features of URLs are extracted, and then different ML models are trained. The limitation of this approach is that it requires manual feature engineering and it does not consider the sequential patterns in the URL. Therefore, deep learning (DL) models are used to solve these issues since they are able to perform featureless detection. Furthermore, DL models give better accuracy and generalization to newly designed URLs; however, the results of our study show that these models, such as any other DL models, can be susceptible to adversarial attacks. In this paper, we examine the robustness of these models and demonstrate the importance of considering this susceptibility before applying such detection systems in real-world solutions. We propose and demonstrate a black-box attack based on scoring functions with greedy search for the minimum number of perturbations leading to a misclassification. The attack is examined against different types of convolutional neural networks (CNN)-based URL classifiers and it causes a tangible decrease in the accuracy with more than 56% reduction in the accuracy of the best classifier (among the selected classifiers for this work). Moreover, adversarial training shows promising results in reducing the influence of the attack on the robustness of the model to less than 7% on average.
引用
收藏
页码:921 / 939
页数:19
相关论文
共 30 条
[11]  
Ebrahimi J, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, P31
[12]   Black-box Generation of Adversarial Text Sequences to Evade Deep Learning Classifiers [J].
Gao, Ji ;
Lanchantin, Jack ;
Soffa, Mary Lou ;
Qi, Yanjun .
2018 IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2018), 2018, :50-56
[13]  
Goodfellow I., 2015, PROC INT C LEARN REP
[14]   Machine Learning in IoT Security: Current Solutions and Future Challenges [J].
Hussain, Fatima ;
Hussain, Rasheed ;
Hassan, Syed Ali ;
Hossain, Ekram .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2020, 22 (03) :1686-1721
[15]  
Keliris A, 2016, INT TEST CONF P
[16]  
Kuhrer Marc, 2014, Research in Attacks, Intrusions and Defenses. 17th International Symposium (RAID 2014). Proceedings: LNCS 8688, P1, DOI 10.1007/978-3-319-11379-1_1
[17]  
Le H, 2018, URLNET LEARNING URL
[18]  
LLC OpenDNS, 2016, PHISHTANK ANT SIT
[19]  
Ma J, 2009, KDD-09: 15TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, P1245
[20]  
Namatevs I., 2017, Inform. Technol. Manage. Sci., V20, P40, DOI DOI 10.1515/ITMS-2017-0007