Explaining poor performance of text-based machine learning models for vulnerability detection

被引:0
作者
Napier, Kollin [1 ]
Bhowmik, Tanmay [2 ]
Chen, Zhiqian [2 ]
机构
[1] Mississippi Gulf Coast Community Coll, Mississippi Artificial Intelligence Network MAIN, Perkinston, MS 39507 USA
[2] Mississippi State Univ, Dept Comp Sci & Engn, Mississippi State, MS USA
关键词
Text-based analysis; Machine learning models; Explainability;
D O I
10.1007/s10664-024-10519-8
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
With an increase of severity in software vulnerabilities, machine learning models are being adopted to combat this threat. Given the possibilities towards usage of such models, research in this area has introduced various approaches. Although models may differ in performance, there is an overall lack of explainability in understanding how a model learns and predicts. Furthermore, recent research suggests that models perform poorly in detecting vulnerabilities when interpreting source code as text, known as "text-based" models. To help explain this poor performance, we explore the dimensions of explainability. From recent studies on text-based models, we experiment with removal of overlapping features present in training and testing datasets, deemed "cross-cutting". We conduct scenario experiments removing such "cross-cutting" data and reassessing model performance. Based on the results, we examine how removal of these "cross-cutting" features may affect model performance. Our results show that removal of "cross-cutting" features may provide greater performance of models in general, thus leading to explainable dimensions regarding data dependency and agnostic models. Overall, we conclude that model performance can be improved, and explainable aspects of such models can be identified via empirical analysis of the models' performance.
引用
收藏
页数:44
相关论文
共 52 条
[21]   SySeVR: A Framework for Using Deep Learning to Detect Software Vulnerabilities [J].
Li, Zhen ;
Zou, Deqing ;
Xu, Shouhuai ;
Jin, Hai ;
Zhu, Yawei ;
Chen, Zhaoxuan .
IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2022, 19 (04) :2244-2258
[22]  
Li Z, 2018, Arxiv, DOI arXiv:1801.01681
[23]   Software Vulnerability Detection Using Deep Neural Networks: A Survey [J].
Lin, Guanjun ;
Wen, Sheng ;
Han, Qing-Long ;
Zhang, Jun ;
Xiang, Yang .
PROCEEDINGS OF THE IEEE, 2020, 108 (10) :1825-1848
[24]   Software Vulnerability Discovery via Learning Multi-Domain Knowledge Bases [J].
Lin, Guanjun ;
Zhang, Jun ;
Luo, Wei ;
Pan, Lei ;
De Vel, Olivier ;
Montague, Paul ;
Xiang, Yang .
IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2021, 18 (05) :2469-2485
[25]   POSTER: Vulnerability Discovery with Function Representation Learning from Unlabeled Projects [J].
Lin, Guanjun ;
Zhang, Jun ;
Luo, Wei ;
Pan, Lei ;
Xiang, Yang .
CCS'17: PROCEEDINGS OF THE 2017 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2017, :2539-2541
[26]   The Mythos of Model Interpretability [J].
Lipton, Zachary C. .
COMMUNICATIONS OF THE ACM, 2018, 61 (10) :36-43
[27]   Software Vulnerability Discovery Techniques: A Survey [J].
Liu, Bingchang ;
Shi, Liang ;
Cai, Zhuhua ;
Li, Min .
2012 FOURTH INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION NETWORKING AND SECURITY (MINES 2012), 2012, :152-156
[28]   CD-VulD: Cross-Domain Vulnerability Discovery Based on Deep Domain Adaptation [J].
Liu, Shigang ;
Lin, Guanjun ;
Qu, Lizhen ;
Zhang, Jun ;
De Vel, Olivier ;
Montague, Paul ;
Xiang, Yang .
IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2022, 19 (01) :438-451
[29]   What Types of Defects Are Really Discovered in Code Reviews? [J].
Mantyla, Mika V. ;
Lassenius, Casper .
IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2009, 35 (03) :430-448
[30]  
Molnar C., 2018, A guide for making black box models explainable