Patch-Wise Attack for Fooling Deep Neural Network

被引:90
作者
Gao, Lianli [1 ]
Zhang, Qilong [1 ]
Song, Jingkuan [1 ]
Liu, Xianglong [2 ]
Shen, Heng Tao [1 ]
机构
[1] Univ Elect Sci & Technol China, Ctr Future Media & Sch Comp Sci & Engn, Chengdu, Peoples R China
[2] Beihang Univ, Beijing, Peoples R China
来源
COMPUTER VISION - ECCV 2020, PT XXVIII | 2020年 / 12373卷
基金
中国国家自然科学基金;
关键词
Adversarial examples; Patch-wise; Black-box attack; Transferability;
D O I
10.1007/978-3-030-58604-1_19
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
By adding human-imperceptible noise to clean images, the resultant adversarial examples can fool other unknown models. Features of a pixel extracted by deep neural networks (DNNs) are influenced by its surrounding regions, and different DNNs generally focus on different discriminative regions in recognition. Motivated by this, we propose a patch-wise iterative algorithm - a black-box attack towards mainstream normally trained and defense models, which differs from the existing attack methods manipulating pixel-wise noise. In this way, without sacrificing the performance of white-box attack, our adversarial examples can have strong transferability. Specifically, we introduce an amplification factor to the step size in each iteration, and one pixel's overall gradient overflowing the epsilon-constraint is properly assigned to its surrounding regions by a project kernel. Our method can be generally integrated to any gradient-based attack methods. Compared with the current state-of-the-art attacks, we significantly improve the success rate by 9.2% for defense models and 3.7% for normally trained models on average. Our code is available at https://github.com/qilong-zhang/Patch-wise-iterative-attack
引用
收藏
页码:307 / 322
页数:16
相关论文
共 37 条
[1]  
[Anonymous], 2017, ICLR
[2]  
[Anonymous], 2016, arXiv
[3]  
Cisse M., 2017, CoRR abs/1707.05373
[4]   Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks [J].
Dong, Yinpeng ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4307-4316
[5]   Boosting Adversarial Attacks with Momentum [J].
Dong, Yinpeng ;
Liao, Fangzhou ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun ;
Hu, Xiaolin ;
Li, Jianguo .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9185-9193
[6]  
Efros AA, 2001, COMP GRAPH, P341, DOI 10.1145/383259.383296
[7]   Robust Physical-World Attacks on Deep Learning Visual Classification [J].
Eykholt, Kevin ;
Evtimov, Ivan ;
Fernandes, Earlence ;
Li, Bo ;
Rahmati, Amir ;
Xiao, Chaowei ;
Prakash, Atul ;
Kohno, Tadayoshi ;
Song, Dawn .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :1625-1634
[8]  
Goodfellow IJ., 2014, CORR
[9]  
Guo C., 2018, Countering Adversarial Images using Input Transformations, DOI DOI 10.1109/GLOCOM.2018.8647866
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778