Improving the robustness of adversarial attacks using an affine-invariant gradient estimator

被引:4
作者
Xiang, Wenzhao [1 ,4 ]
Su, Hang [2 ,3 ]
Liu, Chang [1 ]
Guo, Yandong [5 ]
Zheng, Shibao [1 ]
机构
[1] Shanghai Jiao Tong Univ, Inst Image Commun & Networks Engn, Dept Elect Engn EE, Shanghai 200240, Peoples R China
[2] Tsinghua Univ, Inst Artificial Intelligence, Tsinghua Bosch Joint Ctr Machine Learning, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[3] Zhongguancun Lab, Beijing 100080, Peoples R China
[4] Pengcheng Lab, Shenzhen 518055, Peoples R China
[5] OPPO Res Inst, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Adversarial attack; Deep neural networks; Affine invariance; Transferability;
D O I
10.1016/j.cviu.2023.103647
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As designers of artificial intelligence try to outwit hackers, both sides continue to hone in on AI's inherent vulnerabilities. Designed and trained from certain statistical distributions of data, deep neural networks (DNNs) remain vulnerable to deceptive inputs that violate a DNN's statistical, predictive assumptions. Before being fed into a neural network, however, most existing adversarial examples cannot maintain malicious functionality when applied to an affine transformation. For practical purposes, maintaining that malicious functionality serves as an important measure of the robustness of adversarial attacks. To help DNNs learn to defend themselves more thoroughly against attacks, we propose an affine-invariant adversarial attack, which can consistently produce more robust adversarial examples over affine transformations. For efficiency, we propose to disentangle current affine-transformation strategies from the Euclidean geometry coordinate plane with its geometric translations, rotations and dilations; we reformulate the latter two in polar coordinates. Afterwards, we construct an affine-invariant gradient estimator by convolving the gradient at the original image with derived kernels, which can be integrated with any gradient-based attack methods. Extensive experiments on ImageNet, including some experiments under physical condition, demonstrate that our method can significantly improve the affine invariance of adversarial examples and, as a byproduct, improve the transferability of adversarial examples, compared with alternative state-of-the-art methods.
引用
收藏
页数:11
相关论文
共 47 条
[1]  
Al-Qizwini M, 2017, IEEE INT VEH SYM, P89, DOI 10.1109/IVS.2017.7995703
[2]  
Allen-Zhu Z, 2019, PR MACH LEARN RES, V97
[3]  
Athalye A, 2018, PR MACH LEARN RES, V80
[4]  
Biggio Battista, 2013, Machine Learning and Knowledge Discovery in Databases. European Conference, ECML PKDD 2013. Proceedings: LNCS 8190, P387, DOI 10.1007/978-3-642-40994-3_25
[5]  
Brendel W, 2018, Arxiv, DOI arXiv:1712.04248
[6]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[7]  
Cheng SY, 2019, ADV NEUR IN, V32
[8]   Benchmarking Adversarial Robustness on Image Classification [J].
Dong, Yinpeng ;
Fu, Qi-An ;
Yang, Xiao ;
Pang, Tianyu ;
Su, Hang ;
Xiao, Zihao ;
Zhu, Jun .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :318-328
[9]   Efficient Decision-based Black-box Adversarial Attacks on Face Recognition [J].
Dong, Yinpeng ;
Su, Hang ;
Wu, Baoyuan ;
Li, Zhifeng ;
Liu, Wei ;
Zhang, Tong ;
Zhu, Jun .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :7706-7714
[10]   Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks [J].
Dong, Yinpeng ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4307-4316