GradFuzz: Fuzzing deep neural networks with gradient vector coverage for adversarial examples

被引:5
|
作者
Park, Leo Hyun [1 ]
Chung, Soochang [1 ]
Kim, Jaeuk [1 ]
Kwon, Taekyoung [1 ]
机构
[1] Yonsei Univ, Grad Sch Informat, Informat Secur Lab, Seoul 03722, South Korea
基金
新加坡国家研究基金会;
关键词
Deep learning security; Coverage -guided DNN fuzzing; Gradient vector coverage;
D O I
10.1016/j.neucom.2022.12.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are susceptible to adversarial attacks that add perturbations to the input data, leading to misclassification errors and causing machine-learning systems to fail. For defense, adver-sarial training leverages possible crashing inputs, i.e., adversarial examples; but, the input space of DNNs is enormous and high-dimensional, making it difficult to find in a wide range. Coverage-guided fuzzing is promising in this respect. However, this leaves the question of what coverage metrics are appropriate for DNNs. We observed that the abilities of existing coverage metrics are limited. They lack gradual guidance toward crashes because of a simple search for a wide neuron activation area. None of the existing approaches can simultaneously achieve high crash quantity, high crash diversity, and efficient fuzzing time. Apart from this, the evaluation methodologies adopted by state-of-the-art fuzzers need rigorous improvements. To address these problems, we present a new DNN fuzzer named GradFuzz. Our idea is the gradient vector coverage, which provides gradual guidance to misclassified categories. We imple-mented our system and performed experiments under rigorous evaluation methodologies. Our evalua-tion results indicate that GradFuzz outperforms state-of-the-art DNN fuzzers: GradFuzz can locate a more diverse set of errors, beneficial to adversarial training, on the MNIST and CIFAR-10 datasets without sacrificing both crash quantity and fuzzing efficiency. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:165 / 180
页数:16
相关论文
共 50 条
  • [31] Detecting Adversarial Examples for Deep Neural Networks via Layer Directed Discriminative Noise Injection
    Wang, Si
    Liu, Wenye
    Chang, Chip-Hong
    PROCEEDINGS OF THE 2019 ASIAN HARDWARE ORIENTED SECURITY AND TRUST SYMPOSIUM (ASIANHOST), 2019,
  • [32] Toward deep neural networks robust to adversarial examples, using augmented data importance perception
    Chen, Zhiming
    Xue, Wei
    Tian, Weiwei
    Wu, Yunhua
    Hua, Bing
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [33] Crafting adversarial example with adaptive root mean square gradient on deep neural networks
    Xiao, Yatie
    Pun, Chi-Man
    Liu, Bo
    NEUROCOMPUTING, 2020, 389 : 179 - 195
  • [34] AccelAT: A Framework for Accelerating the Adversarial Training of Deep Neural Networks Through Accuracy Gradient
    Nikfam, Farzad
    Marchisio, Alberto
    Martina, Maurizio
    Shafique, Muhammad
    IEEE Access, 2022, 10 : 108997 - 109007
  • [35] AccelAT: A Framework for Accelerating the Adversarial Training of Deep Neural Networks Through Accuracy Gradient
    Nikfam, Farzad
    Marchisio, Alberto
    Martina, Maurizio
    Shafique, Muhammad
    IEEE ACCESS, 2022, 10 : 108997 - 109007
  • [36] Exploring adversarial examples and adversarial robustness of convolutional neural networks by mutual information
    Zhang J.
    Qian W.
    Cao J.
    Xu D.
    Neural Computing and Applications, 2024, 36 (23) : 14379 - 14394
  • [37] Adversarial Examples Against Deep Neural Network based Steganalysis
    Zhang, Yiwei
    Zhang, Weiming
    Chen, Kejiang
    Liu, Jiayang
    Liu, Yujia
    Yu, Nenghai
    PROCEEDINGS OF THE 6TH ACM WORKSHOP ON INFORMATION HIDING AND MULTIMEDIA SECURITY (IH&MMSEC'18), 2018, : 67 - 72
  • [38] Enhancing Adversarial Examples on Deep Q Networks with Previous Information
    Sooksatra, Korn
    Rivas, Pablo
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [39] Pruning Adversarially Robust Neural Networks without Adversarial Examples
    Jian, Tong
    Wang, Zifeng
    Wang, Yanzhi
    Dy, Jennifer
    Ioannidis, Stratis
    2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2022, : 993 - 998
  • [40] Examining the Proximity of Adversarial Examples to Class Manifolds in Deep Networks
    Pocos, Stefan
    Beckova, Iveta
    Farkas, Igor
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT IV, 2022, 13532 : 645 - 656