Adversarial Robustness of Deep Learning-Based Malware Detectors via (De)Randomized Smoothing

被引:2
作者
Gibert, Daniel [1 ]
Zizzo, Giulio [2 ]
Le, Quan [1 ]
Planes, Jordi [3 ]
机构
[1] Univ Coll Dublin, CeADAR, Dublin D04 V2N9, Ireland
[2] IBM Res Europe, Dublin D15 HN66, Ireland
[3] Univ Lleida, Dept Comp Engn & Digital Design, Lleida, Spain
基金
欧盟地平线“2020”;
关键词
Adversarial defense; (de)randomized smoothing; evasion attacks; machine learning; malware detection;
D O I
10.1109/ACCESS.2024.3392391
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning-based malware detectors have been shown to be susceptible to adversarial malware examples, i.e. malware examples that have been deliberately manipulated in order to avoid detection. In light of the vulnerability of deep learning detectors to subtle input file modifications, we propose a practical defense against adversarial malware examples inspired by (de)randomized smoothing. In this work, we reduce the chances of sampling adversarial content injected by malware authors by selecting correlated subsets of bytes, rather than using Gaussian noise to randomize inputs like in the Computer Vision domain. During training, our chunk-based smoothing scheme trains a base classifier to make classifications on a subset of contiguous bytes or chunk of bytes. At test time, a large number of chunks are then classified by a base classifier and the consensus among these classifications is then reported as the final prediction. We propose two strategies to determine the location of the chunks used for classification: 1) randomly selecting the locations of the chunks and 2) selecting contiguous adjacent chunks. To showcase the effectiveness of our approach, we have trained two classifiers with our chunk-based smoothing schemes on the BODMAS dataset. Our findings reveal that the chunk-based smoothing classifiers exhibit greater resilience against adversarial malware examples generated with state-of-the-art evasion attacks, outperforming a non-smoothed classifier and a randomized smoothing-based classifier by a great margin.
引用
收藏
页码:61152 / 61162
页数:11
相关论文
共 26 条
[1]   Malware Detection by Eating a Whole APK [J].
Al-Fawa'reh, Mohammad ;
Saif, Amal ;
Jafar, Mousa Tayseer ;
Elhassan, Ammar .
INTERNATIONAL CONFERENCE FOR INTERNET TECHNOLOGY AND SECURED TRANSACTIONS (ICITST-2020), 2020, :107-113
[2]  
Biggio Battista, 2013, Machine Learning and Knowledge Discovery in Databases, P387, DOI [10.1007/978-3-642-40994-3_25, DOI 10.1007/978-3-642-40994-3_25, DOI 10.1007/978-3-642-40994-3]
[3]   Learning Understandable Neural Networks With Nonnegative Weight Constraints [J].
Chorowski, Jan ;
Zurada, Jacek M. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (01) :62-69
[4]  
Cohen J, 2019, PR MACH LEARN RES, V97
[5]  
Demetrio L, 2024, Arxiv, DOI arXiv:2104.12848
[6]  
Demetrio L, 2019, Arxiv, DOI arXiv:1901.03583
[7]   Adversarial EXEmples: A Survey and Experimental Evaluation of Practical Attacks on Machine Learning for Windows Malware Detection [J].
Demetrio, Luca ;
Coull, Scott E. ;
Biggio, Battista ;
Lagorio, Giovanni ;
Armando, Alessandro ;
Roli, Fabio .
ACM TRANSACTIONS ON PRIVACY AND SECURITY, 2021, 24 (04)
[8]   Functionality-Preserving Black-Box Optimization of Adversarial Windows Malware [J].
Demetrio, Luca ;
Biggio, Battista ;
Lagorio, Giovanni ;
Roli, Fabio ;
Armando, Alessandro .
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 :3469-3478
[9]  
Fleshman W, 2019, Arxiv, DOI arXiv:1806.06108
[10]   A Wolf in Sheep's Clothing: Query-Free Evasion Attacks Against Machine Learning-Based Malware Detectors with Generative Adversarial Networks [J].
Gibcrt, Daniel ;
Planes, Jordi ;
Lc, Quan ;
Zizzo, Giulio .
2023 IEEE EUROPEAN SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS, EUROS&PW, 2023, :415-426