Improving transferable adversarial attack for vision transformers via global attention and local drop

被引:0
作者
Tuo Li
Yahong Han
机构
[1] Tianjin University,College of Intelligence and Computing
来源
Multimedia Systems | 2023年 / 29卷
关键词
Adversarial examples; Vision transformer; Transferability; Self-attention;
D O I
暂无
中图分类号
学科分类号
摘要
Vision Transformers (ViTs) have been a new paradigm in several computer vision tasks, yet they are susceptible to adversarial examples. Recent studies show it is difficult to transfer adversarial examples generated by ViTs to other models. Existing methods have poor transferability because they do not target the specific structural characteristics (e.g., self-attention and patch-embedding) of ViTs. To address this problem and further boost transferability, we propose a method, namely Global Attention and Local Drop (GALD), to boost the transferability of adversarial examples from ViTs to other models, including ViTs and convolutional neural networks (CNNs). Specifically, our method contains two parts: Global Attention Guidance (GAG) and Drop Patch (DP). The GAG improves the attention representation in shallow layers by adding global guidance attention to every layer except the final layer of ViTs. Therefore, the perturbations could focus on the object regions. DP randomly drops some patches in every iteration to diversify the input patterns and mitigate overfitting of adversarial examples to the surrogate model. Experiments show that adversarial examples generated by our method own the best transferability to black-box models with unknown structures. Code is available at Link.
引用
收藏
页码:3467 / 3480
页数:13
相关论文
共 32 条
[1]  
Bai Y(2021)Are transformers more robust than CNNs? Ad Neural Inf Process Syst 34 26831-26843
[2]  
Mei J(2022)Toward understanding and boosting adversarial transferability from a distribution perspective IEEE Trans. Image Process. 31 6487-6501
[3]  
Yuille AL(2021)Intriguing properties of vision transformers Adv. Neural Inf. Process. Syst. 34 23296-23308
[4]  
Xie C(2015)ImageNet large scale visual recognition challenge Int. J. Comput. Vis. 115 211-252
[5]  
Zhu Y(2021)On success and simplicity: a second look at transferable targeted attacks Adv. Neural Inf. Process. Syst. 34 6115-6128
[6]  
Chen Y(undefined)undefined undefined undefined undefined-undefined
[7]  
Li X(undefined)undefined undefined undefined undefined-undefined
[8]  
Chen K(undefined)undefined undefined undefined undefined-undefined
[9]  
He Y(undefined)undefined undefined undefined undefined-undefined
[10]  
Tian X(undefined)undefined undefined undefined undefined-undefined