Evading Machine Learning Botnet Detection Models via Deep Reinforcement Learning

被引:33
作者
Wu, Di [1 ,2 ]
Fang, Binxing [3 ,4 ,5 ]
Wang, Junnan [1 ,2 ]
Liu, Qixu [1 ,2 ]
Cui, Xiang [1 ,3 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Guangzhou Univ, Cyberspace Inst Adv Technol, Guangzhou, Guangdong, Peoples R China
[4] UESTC Guangdong, Inst Elect & Informat Engn, Dongguan, Guangdong, Peoples R China
[5] Beijing Univ Posts & Telecommun, Sch Cyberspace Secur, Beijing, Peoples R China
来源
ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC) | 2019年
关键词
botnet; adversarial; reinforcement learning;
D O I
10.1109/icc.2019.8761337
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Botnets are one of predominant threats to Internet security. To date, machine learning technology has wide application in botnet detection because that it is able to summarize the features of existing attacks and generalize to never-before-seen botnet families. However, recent works in adversarial machine learning have shown that attackers are able to bypass the detection model by constructing specific samples, which due to many algorithms are vulnerable to almost imperceptible perturbations of their inputs. According to the degree of adversaries' knowledge about the model, adversarial attacks can be classified into several groups, such as gradient- and score-based attacks. In this paper, we propose a more general framework based on deep reinforcement learning (DRL), which effectively generates adversarial traffic flows to deceive the detection model by automatically adding perturbations to samples. Throughout the process, the target detector will be regarded as a black box and more close to realistic attack circumstance. A reinforcement learning agent is equipped for updating the adversarial samples by combining the feedback from the target model (i.e. benign or malicious) and the sequence of actions, which is able to change the temporal and spatial features of the traffic flows while maintaining the original functionality and executability. The experiment results show that the evasion rates of adversarial botnet flows are significantly improved. Furthermore, with the perspective of defense, this research can help the detection model spot its defect and thus enhance the robustness.
引用
收藏
页数:6
相关论文
共 25 条
[1]   DeepDGA: Adversarially-Tuned Domain Generation and Detection [J].
Anderson, Hyrum S. ;
Woodbridge, Jonathan ;
Filar, Bobby .
AISEC'16: PROCEEDINGS OF THE 2016 ACM WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, 2016, :13-21
[2]  
[Anonymous], ADV NEURAL INFORM PR
[3]  
[Anonymous], 2016, P 2016 IEEE BIENNIAL, DOI DOI 10.1109/ARGENCON.2016.7585247
[4]  
[Anonymous], 2008, BOTMINER CLUSTERING
[5]  
[Anonymous], 2017, P 2017 ACM AS C COMP
[6]  
[Anonymous], 2016, P 2016 NETWORK DISTR
[7]  
Bilge L, 2012, 28TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE (ACSAC 2012), P129
[8]  
Brendel Wieland, 2017, arXiv:1712.04248.
[9]  
Carlini N, 2016, PROCEEDINGS OF THE 25TH USENIX SECURITY SYMPOSIUM, P513
[10]  
Cheung V, 2016, OPENAI GYM