Black-box adversarial sample generation based on differential evolution

被引:31
作者
Lin, Junyu [1 ,2 ]
Xu, Lei [1 ,2 ]
Liu, Yingqi [3 ]
Zhang, Xiangyu [3 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
[2] Nanjing Univ, Dept Comp Sci & Technol, Nanjing, Peoples R China
[3] Purdue Univ, Dept Comp Sci, W Lafayette, IN 47907 USA
关键词
Adversarial samples; Differential evolution; Black-box testing; Deep Neural Network;
D O I
10.1016/j.jss.2020.110767
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep Neural Networks (DNNs) are being used in various daily tasks such as object detection, speech processing, and machine translation. However, it is known that DNNs suffer from robustness problems - perturbed inputs called adversarial samples leading to misbehaviors of DNNs. In this paper, we propose a black-box technique called Black-box Momentum Iterative Fast Gradient Sign Method (BMI-FGSM) to test the robustness of DNN models. The technique does not require any knowledge of the structure or weights of the target DNN. Compared to existing white-box testing techniques that require accessing model internal information such as gradients, our technique approximates gradients through Differential Evolution and uses approximated gradients to construct adversarial samples. Experimental results show that our technique can achieve 100% success in generating adversarial samples to trigger misclassification, and over 95% success in generating samples to trigger misclassification to a specific target output label. It also demonstrates better perturbation distance and better transferability. Compared to the state-of-the-art black-box technique, our technique is more efficient. Furthermore, we conduct testing on the commercial Aliyun API and successfully trigger its misbehavior within a limited number of queries, demonstrating the feasibility of real-world black-box attack. (C) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页数:11
相关论文
共 48 条
[21]  
Kurakin A., 2017, INT C LEARN REPR
[22]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324
[23]  
Lin M, 2014, PUBLIC HEALTH NUTR, V17, P2029, DOI [10.1109/PLASMA.2013.6634954, 10.1017/S1368980013002176]
[24]   A survey on deep learning in medical image analysis [J].
Litjens, Geert ;
Kooi, Thijs ;
Bejnordi, Babak Ehteshami ;
Setio, Arnaud Arindra Adiyoso ;
Ciompi, Francesco ;
Ghafoorian, Mohsen ;
van der Laak, Jeroen A. W. M. ;
van Ginneken, Bram ;
Sanchez, Clara I. .
MEDICAL IMAGE ANALYSIS, 2017, 42 :60-88
[25]   DeepGauge: Multi-Granularity Testing Criteria for Deep Learning Systems [J].
Ma, Lei ;
Juefei-Xu, Felix ;
Zhang, Fuyuan ;
Sun, Jiyuan ;
Xue, Minhui ;
Li, Bo ;
Chen, Chunyang ;
Su, Ting ;
Li, Li ;
Liu, Yang ;
Zhao, Jianjun ;
Wang, Yadong .
PROCEEDINGS OF THE 2018 33RD IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMTED SOFTWARE ENGINEERING (ASE' 18), 2018, :120-131
[26]  
Ma L, 2019, 2019 IEEE 26TH INTERNATIONAL CONFERENCE ON SOFTWARE ANALYSIS, EVOLUTION AND REENGINEERING (SANER), P614, DOI [10.1109/saner.2019.8668044, 10.1109/SANER.2019.8668044]
[27]   DeepMutation: Mutation Testing of Deep Learning Systems [J].
Ma, Lei ;
Zhang, Fuyuan ;
Sun, Jiyuan ;
Xue, Minhui ;
Li, Bo ;
Juefei-Xu, Felix ;
Xie, Chao ;
Li, Li ;
Liu, Yang ;
Zhao, Jianjun ;
Wang, Yadong .
2018 29TH IEEE INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING (ISSRE), 2018, :100-111
[28]  
Madry A., 2017, ARXIV
[29]   Universal adversarial perturbations [J].
Moosavi-Dezfooli, Seyed-Mohsen ;
Fawzi, Alhussein ;
Fawzi, Omar ;
Frossard, Pascal .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :86-94
[30]   Simple Black-Box Adversarial Attacks on Deep Neural Networks [J].
Narodytska, Nina ;
Kasiviswanathan, Shiva .
2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, :1310-1318