Prediction of Vulnerability Characteristics Based on Vulnerability Description and Prompt Learning

被引:0
作者
Li, Xiangwei [1 ]
Ren, Xiaoning [1 ]
Xue, Yinxing [1 ]
Xing, Zhenchang [2 ,3 ]
Sun, Jiamou [2 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] CSIRO, Data61, Canberra, ACT, Australia
[3] Australian Natl Univ, Canberra, ACT, Australia
来源
2023 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE ANALYSIS, EVOLUTION AND REENGINEERING, SANER | 2023年
关键词
vulnerability characteristics; vulnerability assessment; prompt learning; pre-trained language model;
D O I
10.1109/SANER56733.2023.00062
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Identifying which vulnerabilities need to be prioritized is a long-term challenge in IT security, especially as the number of vulnerabilities grows. Faced with a large number of vulnerability reports, there is an urgent need for automated tools or models to assess the potential severity and exploitability of vulnerabilities. This will help security experts screen vulnerabilities that should be focused on. In this study, we aim to predict vulnerability severity and exploitability characteristics using only vulnerability descriptions. Some previous studies are based on traditional deep learning models, and their performance is relatively backward in the current era of pre-trained language models (PLMs). Therefore, we introduce a prompt learning method based on PLMs to predict vulnerability characteristics. The conventional fine-tuning PLMs method is difficult to make full use of the domain knowledge in PLMs and performs poorly with less training data. Unlike the fine-tuning paradigm, prompt learning imitates the pre-training process of PLM by reconstructing the task input and adding prompts, and uses the output of PLM itself as the prediction output. Combined with prompt ensembling and transfer learning, the performance of prompt learning in the above tasks is further improved. Our experiments show that prompt learning can make more effective use of the knowledge in PLMs. Compared with fine-tuning PLMs and other deep learning models, prompt learning based on BERT or RoBERTa achieves better performance in the above tasks. This advantage is more significant in predicting exploitability with few samples, which proves the ability of prompt learning in few-sample scenarios. In addition, prompt learning also shows the transferability between different tasks in the domain.
引用
收藏
页码:604 / 615
页数:12
相关论文
共 49 条
  • [1] Brown TB, 2020, ADV NEUR IN, V33
  • [2] CISA, ABOUT US
  • [3] CISA, BIND OP DIR 22 01
  • [4] CISA, KNOWN EXPL VULN CAT
  • [5] Das S.S., 2021, PROC IEEE 8 INT C DA, P1
  • [6] Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
  • [7] Ding N, 2021, Arxiv, DOI arXiv:2111.01998
  • [8] Dodge J, 2020, Arxiv, DOI arXiv:2002.06305
  • [9] Feng ZY, 2020, Arxiv, DOI [arXiv:2002.08155, DOI 10.48550/ARXIV.2002.08155, 10.48550/arXiv.2002.08155]
  • [10] Catastrophic forgetting in connectionist networks
    French, RM
    [J]. TRENDS IN COGNITIVE SCIENCES, 1999, 3 (04) : 128 - 135