Gradient Aggregation Boosting Adversarial Examples Transferability Method

被引:0
作者
Deng, Shiyun [1 ]
Ling, Jie [1 ]
机构
[1] School of Computer, Guangdong University of Technology, Guangzhou
关键词
adversarial attacks; deep neural network; gradient aggregation; transferability;
D O I
10.3778/j.issn.1002-8331.2304-0174
中图分类号
学科分类号
摘要
Image classification models based on deep neural networks are vulnerable to adversarial examples. Existing studies have shown that white-box attacks have been able to achieve a high attack success rate, but the transferability of adversarial examples is low when attacking other models. In order to improve the transferability of adversarial attacks, this paper proposes a gradient aggregation method to enhance the transferability of adversarial examples. Firstly, the original image is mixed with other class images in a specific ratio to obtain a mixed image. By comprehensively considering the information of different categories of images and balancing the gradient contributions between categories, the influence of local oscillations can be avoided. Secondly, in the iterative process, the gradient information of other data points in the neighborhood of the current point is aggregated to optimize the gradient direction, avoiding excessive dependence on a single data point, and thus generating adversarial examples with stronger mobility. Experimental results on the ImageNet dataset show that the proposed method significantly improves the success rate of black-box attacks and the transferability of adversarial examples. On the single-model attack, the average attack success rate of the method in this paper is 88.5% in the four conventional training models, which is 2.7 percentage points higher than the Admix method; the average attack success rate on the integrated model attack reaches 92.7%. In addition, the proposed method can be integrated with the transformation-based adversarial attack method, and the average attack success rate on the three adversarial training models is 10.1 percentage points, higher than that of the Admix method, which enhances the transferability of adversarial attacks. © 2024 Journal of Computer Engineering and Applications Beijing Co., Ltd.; Science Press. All rights reserved.
引用
收藏
页码:275 / 282
页数:7
相关论文
共 20 条
[1]  
LI H, HUANG H, CHEN L, Et al., Adversarial examples for CNN- based SAR image classification: an experience study, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 14, pp. 1333-1347, (2020)
[2]  
ROY A M, BOSE R, BHADURI J., A fast accurate fine-grain object detection model based on YOLOv4 deep neural network, Neural Computing and Applications, 34, pp. 3895-3921, (2022)
[3]  
LI L, MU X, LI S, Et al., A review of face recognition technology, IEEE Access, 8, pp. 139110-139120, (2020)
[4]  
GOODFELLOW I J, SHLENS J, SZEGEDY C., Explaining and harnessing adversarial examples, (2014)
[5]  
PAPERNOT N, MCDANIEL P, JHA S, Et al., The limitations of deep learning in adversarial settings, Proceedings of the IEEE European Symposium on Security and Privacy, pp. 372-387, (2016)
[6]  
ZHENG H, ZHANG Z, GU J, Et al., Efficient adversarial training with transferable adversarial examples, Proceedings of Conference on Computer Vision and Pattern Recognition, pp. 1181-1190, (2020)
[7]  
WANG X, HE X, WANG J, Et al., Admix: enhancing the transferability of adversarial attacks, Proceedings of the International Conference on Computer Vision, pp. 16158-16167, (2021)
[8]  
DEMONTIS A, MELIS M, PINTOR M, Et al., Why do adversarial attacks transfer? explaining transferability of evasion and poisoning attacks, Proceedings of the 28th USENIX Security Symposium, pp. 321-338, (2019)
[9]  
WANG Z, GUO H, ZHANG Z, Et al., Feature importance-aware transferable adversarial attacks, Proceedings of the International Conference on Computer Vision, pp. 7639-7648, (2021)
[10]  
YANG K Y, YAU J H, LI F F, Et al., A study of face obfuscation in imagenet, Proceedings of the 39th International Conference on Machine Learning, pp. 25313-25330, (2022)