Practical Black-Box Attacks on Deep Neural Networks Using Efficient Query Mechanisms

被引:145
作者
Bhagoji, Arjun Nitin [1 ,2 ]
He, Warren [2 ]
Li, Bo [3 ]
Song, Dawn [2 ]
机构
[1] Princeton Univ, Princeton, NJ 08544 USA
[2] Univ Calif Berkeley, Berkeley, CA 94720 USA
[3] Univ Illinois, Champaign, IL USA
来源
COMPUTER VISION - ECCV 2018, PT XII | 2018年 / 11216卷
关键词
Deep neural networks; Image classification; Adversarial examples; Black-box attacks;
D O I
10.1007/978-3-030-01258-8_10
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing black-box attacks on deep neural networks (DNNs) have largely focused on transferability, where an adversarial instance generated for a locally trained model can "transfer" to attack other learning models. In this paper, we propose novel Gradient Estimation black-box attacks for adversaries with query access to the target model's class probabilities, which do not rely on transferability. We also propose strategies to decouple the number of queries required to generate each adversarial sample from the dimensionality of the input. An iterative variant of our attack achieves close to 100% attack success rates for both targeted and untargeted attacks on DNNs. We carry out a thorough comparative evaluation of black-box attacks and show that Gradient Estimation attacks achieve attack success rates similar to state-of-the-art white-box attacks on the MNIST and CIFAR-10 datasets. We also apply the Gradient Estimation attacks successfully against real-world classifiers hosted by Clarifai. Further, we evaluate black-box attacks against state-of-the-art defenses based on adversarial training and show that the Gradient Estimation attacks are very effective even against these defenses.
引用
收藏
页码:158 / 174
页数:17
相关论文
共 32 条
[1]  
[Anonymous], 2014, Towards deep neural network architectures robust to adversarial examples
[2]  
[Anonymous], 2016, arXiv
[3]  
[Anonymous], 2017, arXiv
[4]  
Athalye A, 2018, PR MACH LEARN RES, V80
[5]  
Brendel W, 2018, INT C LEARN REPR ICL, DOI DOI 10.48550/ARXIV.1712.04248
[6]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[7]  
Chang E.C, 2017, 24 ACM C COMP COMM S
[8]  
Goodfellow I, 2016, ADAPT COMPUT MACH LE, P1
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]  
Hildebrand F.B., 1962, Advanced calculus for applications, V63