Toward Visual Distortion in Black-Box Attacks

被引:12
作者
Li, Nannan [1 ]
Chen, Zhenzhong [1 ]
机构
[1] Wuhan Univ, Sch Remote Sensing & Informat Engn, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Distortion; Visualization; Measurement; Loss measurement; Optimization; Convergence; Training; Black-box attack; adversarial examples; classification;
D O I
10.1109/TIP.2021.3092822
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Constructing adversarial examples in a black-box threat model injures the original images by introducing visual distortion. In this paper, we propose a novel black-box attack approach that can directly minimize the induced distortion by learning the noise distribution of the adversarial example, assuming only loss-oracle access to the black-box network. To quantify visual distortion, the perceptual distance between the adversarial example and the original image, is introduced in our loss. We first approximate the gradient of the corresponding non-differentiable loss function by sampling noise from the learned noise distribution. Then the distribution is updated using the estimated gradient to reduce visual distortion. The learning continues until an adversarial example is found. We validate the effectiveness of our attack on ImageNet. Our attack results in much lower distortion when compared to the state-of-the-art black-box attacks and achieves 100% success rate on InceptionV3, ResNet50 and VGG16bn. Furthermore, we theoretically prove the convergence of our model. The code is publicly available at https://github.com/Alina-1997/visual-distortion-in-attack.
引用
收藏
页码:6156 / 6167
页数:12
相关论文
共 48 条
[1]   Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey [J].
Akhtar, Naveed ;
Mian, Ajmal .
IEEE ACCESS, 2018, 6 :14410-14430
[2]  
Al-Dujaili A., 2020, P INT C LEARN REPR A
[3]   Square Attack: A Query-Efficient Black-Box Adversarial Attack via Random Search [J].
Andriushchenko, Maksym ;
Croce, Francesco ;
Flammarion, Nicolas ;
Hein, Matthias .
COMPUTER VISION - ECCV 2020, PT XXIII, 2020, 12368 :484-501
[4]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[5]  
Carlini N, 2017, PROCEEDINGS OF THE 10TH ACM WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2017, P3, DOI 10.1145/3128572.3140444
[6]  
Cheng SY, 2019, ADV NEUR IN, V32
[7]  
Gao B., 2017, ARXIV
[8]   Learning to Rank for Blind Image Quality Assessment [J].
Gao, Fei ;
Tao, Dacheng ;
Gao, Xinbo ;
Li, Xuelong .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (10) :2275-2290
[9]  
Goodfellow I. J., 2014, 3 INT C LEARNING REP
[10]   Perceptual quality-preserving black-box attack against deep learning image classifiers [J].
Gragnaniello, Diego ;
Marra, Francesco ;
Verdoliva, Luisa ;
Poggi, Giovanni .
PATTERN RECOGNITION LETTERS, 2021, 147 :142-149