GradFuzz: Fuzzing deep neural networks with gradient vector coverage for adversarial examples

被引:6
作者
Park, Leo Hyun [1 ]
Chung, Soochang [1 ]
Kim, Jaeuk [1 ]
Kwon, Taekyoung [1 ]
机构
[1] Yonsei Univ, Grad Sch Informat, Informat Secur Lab, Seoul 03722, South Korea
基金
新加坡国家研究基金会;
关键词
Deep learning security; Coverage -guided DNN fuzzing; Gradient vector coverage;
D O I
10.1016/j.neucom.2022.12.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are susceptible to adversarial attacks that add perturbations to the input data, leading to misclassification errors and causing machine-learning systems to fail. For defense, adver-sarial training leverages possible crashing inputs, i.e., adversarial examples; but, the input space of DNNs is enormous and high-dimensional, making it difficult to find in a wide range. Coverage-guided fuzzing is promising in this respect. However, this leaves the question of what coverage metrics are appropriate for DNNs. We observed that the abilities of existing coverage metrics are limited. They lack gradual guidance toward crashes because of a simple search for a wide neuron activation area. None of the existing approaches can simultaneously achieve high crash quantity, high crash diversity, and efficient fuzzing time. Apart from this, the evaluation methodologies adopted by state-of-the-art fuzzers need rigorous improvements. To address these problems, we present a new DNN fuzzer named GradFuzz. Our idea is the gradient vector coverage, which provides gradual guidance to misclassified categories. We imple-mented our system and performed experiments under rigorous evaluation methodologies. Our evalua-tion results indicate that GradFuzz outperforms state-of-the-art DNN fuzzers: GradFuzz can locate a more diverse set of errors, beneficial to adversarial training, on the MNIST and CIFAR-10 datasets without sacrificing both crash quantity and fuzzing efficiency. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:165 / 180
页数:16
相关论文
共 47 条
[21]   DEEPSEC: A Uniform Platform for Security Analysis of Deep Learning Model [J].
Ling, Xiang ;
Ji, Shouling ;
Zou, Jiaxu ;
Wang, Jiannan ;
Wu, Chunming ;
Li, Bo ;
Wang, Ting .
2019 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2019), 2019, :673-690
[22]   The Mythos of Model Interpretability [J].
Lipton, Zachary C. .
COMMUNICATIONS OF THE ACM, 2018, 61 (10) :36-43
[23]   DeepGauge: Multi-Granularity Testing Criteria for Deep Learning Systems [J].
Ma, Lei ;
Juefei-Xu, Felix ;
Zhang, Fuyuan ;
Sun, Jiyuan ;
Xue, Minhui ;
Li, Bo ;
Chen, Chunyang ;
Su, Ting ;
Li, Li ;
Liu, Yang ;
Zhao, Jianjun ;
Wang, Yadong .
PROCEEDINGS OF THE 2018 33RD IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMTED SOFTWARE ENGINEERING (ASE' 18), 2018, :120-131
[24]   MODE: Automated Neural Network Model Debugging via State Differential Analysis and Input Selection [J].
Ma, Shiqing ;
Liu, Yingqi ;
Lee, Wen-Chuan ;
Zhang, Xiangyu ;
Grama, Ananth .
ESEC/FSE'18: PROCEEDINGS OF THE 2018 26TH ACM JOINT MEETING ON EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, 2018, :175-186
[25]  
Madry A, 2017, P 6 INT C LEARN REPR
[26]  
Marijan D, 2020, AAAI CONF ARTIF INTE, V34, P13576
[27]   DeepFool: a simple and accurate method to fool deep neural networks [J].
Moosavi-Dezfooli, Seyed-Mohsen ;
Fawzi, Alhussein ;
Frossard, Pascal .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2574-2582
[28]   Scalable Nearest Neighbor Algorithms for High Dimensional Data [J].
Muja, Marius ;
Lowe, David G. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2014, 36 (11) :2227-2240
[29]  
Odena A, 2019, PR MACH LEARN RES, V97
[30]   Practical Black-Box Attacks against Machine Learning [J].
Papernot, Nicolas ;
McDaniel, Patrick ;
Goodfellow, Ian ;
Jha, Somesh ;
Celik, Z. Berkay ;
Swami, Ananthram .
PROCEEDINGS OF THE 2017 ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (ASIA CCS'17), 2017, :506-519