Channel-augmented joint transformation for transferable adversarial attacks

被引:2
作者
Zheng, Desheng [1 ]
Ke, Wuping [1 ]
Li, Xiaoyu [2 ,3 ]
Zhang, Shibin [4 ,5 ]
Yin, Guangqiang [2 ,3 ]
Qian, Weizhong [2 ,3 ]
Zhou, Yong [1 ]
Min, Fan [1 ]
Yang, Shan [6 ]
机构
[1] Southwest Petr Univ, Sch Comp Sci, Chengdu 610500, Peoples R China
[2] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 610054, Peoples R China
[3] Inst Elect & Informat Ind Technol Kash, Silk Rd Talent Bldg,Shenka Ave, Kash 844000, Xinjiang, Peoples R China
[4] Adv Cryptog & Syst Secur Key Lab Sichuan Prov, Chengdu, Peoples R China
[5] Chengdu Univ Informat & Technol, Chengdu, Peoples R China
[6] Jackson State Univ, Dept Chem Phys & Atmospher Sci, Jackson, MS 39217 USA
关键词
Adversarial examples; Neural networks; Computer vision; AI security;
D O I
10.1007/s10489-023-05171-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are vulnerable to adversarial examples that fool the models with tiny perturbations. Although adversarial attacks have achieved incredible attack success rates in the white-box setting, most existing adversaries often exhibit weak transferability in the black-box setting, especially for models with defense mechanisms. In this work, we reveal the cross-model channel redundancy and channel invariance of DNNs and thus propose two channel-augmented methods to improve the transferability of adversarial examples, namely, the channel transformation (CT) method and the channel-invariant Patch (CIP) method. Specifically, channel transformation shuffles and rewrites channels to enhance cross-model feature redundancy in convolution, and channel-invariant patches distinctly weaken different channels to achieve loss-preserving transformation. We compute the aggregated gradients of the transformed dataset to create adversaries with higher transferability. In addition, the two proposed methods can be naturally combined with each other and with almost all other gradient-based methods to further improve performance. Empirical results on the ImageNet dataset demonstrate that our attack methods exhibit higher transferability and achieve higher attack success rates than state-of-the-art gradient-based attacks. Specifically, our attack improves the average attack success rate from 86.9% to 91.0% on normally trained models and from 44.6% to 68.3% on adversarially trained models.
引用
收藏
页码:428 / 442
页数:15
相关论文
共 48 条
[1]  
Cao J., 2022, IEEE Trans. Pattern Anal. Mach, Intell
[2]   Scale-Aware Automatic Augmentations for Object Detection With Dynamic Training [J].
Chen, Yukang ;
Zhang, Peizhen ;
Kong, Tao ;
Li, Yanwei ;
Zhang, Xiangyu ;
Qi, Lu ;
Sun, Jian ;
Jia, Jiaya .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (02) :2367-2383
[3]  
Cohen Jeremy M., 2019, ICML
[4]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[5]   Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks [J].
Dong, Yinpeng ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4307-4316
[6]   Boosting Adversarial Attacks with Momentum [J].
Dong, Yinpeng ;
Liao, Fangzhou ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun ;
Hu, Xiaolin ;
Li, Jianguo .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9185-9193
[7]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[8]  
Gu S, 2019, UMBC STUD COLLECT
[9]  
Guo C., 2018, 6th International Conference on Learning Representations, ICLR 2018-Conference Track Proceedings, International Conference on Learning Representations, DOI DOI 10.1109/GLOCOM.2018.8647866
[10]  
Guo Chuan, 2019, ICML