Channel-augmented joint transformation for transferable adversarial attacks

被引:2
作者
Zheng, Desheng [1 ]
Ke, Wuping [1 ]
Li, Xiaoyu [2 ,3 ]
Zhang, Shibin [4 ,5 ]
Yin, Guangqiang [2 ,3 ]
Qian, Weizhong [2 ,3 ]
Zhou, Yong [1 ]
Min, Fan [1 ]
Yang, Shan [6 ]
机构
[1] Southwest Petr Univ, Sch Comp Sci, Chengdu 610500, Peoples R China
[2] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 610054, Peoples R China
[3] Inst Elect & Informat Ind Technol Kash, Silk Rd Talent Bldg,Shenka Ave, Kash 844000, Xinjiang, Peoples R China
[4] Adv Cryptog & Syst Secur Key Lab Sichuan Prov, Chengdu, Peoples R China
[5] Chengdu Univ Informat & Technol, Chengdu, Peoples R China
[6] Jackson State Univ, Dept Chem Phys & Atmospher Sci, Jackson, MS 39217 USA
关键词
Adversarial examples; Neural networks; Computer vision; AI security;
D O I
10.1007/s10489-023-05171-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are vulnerable to adversarial examples that fool the models with tiny perturbations. Although adversarial attacks have achieved incredible attack success rates in the white-box setting, most existing adversaries often exhibit weak transferability in the black-box setting, especially for models with defense mechanisms. In this work, we reveal the cross-model channel redundancy and channel invariance of DNNs and thus propose two channel-augmented methods to improve the transferability of adversarial examples, namely, the channel transformation (CT) method and the channel-invariant Patch (CIP) method. Specifically, channel transformation shuffles and rewrites channels to enhance cross-model feature redundancy in convolution, and channel-invariant patches distinctly weaken different channels to achieve loss-preserving transformation. We compute the aggregated gradients of the transformed dataset to create adversaries with higher transferability. In addition, the two proposed methods can be naturally combined with each other and with almost all other gradient-based methods to further improve performance. Empirical results on the ImageNet dataset demonstrate that our attack methods exhibit higher transferability and achieve higher attack success rates than state-of-the-art gradient-based attacks. Specifically, our attack improves the average attack success rate from 86.9% to 91.0% on normally trained models and from 44.6% to 68.3% on adversarially trained models.
引用
收藏
页码:428 / 442
页数:15
相关论文
共 48 条
[11]   Minimally Distorted Structured Adversarial Attacks [J].
Kazemi, Ehsan ;
Kerdreux, Thomas ;
Wang, Liqiang .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2023, 131 (01) :160-176
[12]  
Kingma D. P., 2014, P INT C LEARN REPR, P58, DOI DOI 10.48550/ARXIV.1412.6980
[13]  
Kurakin A., 2017, INT C LEARN REPR, DOI DOI 10.48550/ARXIV.1607.02533
[14]   GradDiv: Adversarial Robustness of Randomized Neural Networks via Gradient Diversity Regularization [J].
Lee, Sungyoon ;
Kim, Hoki ;
Lee, Jaewook .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (02) :2645-2651
[15]   Involution: Inverting the Inherence of Convolution for Visual Recognition [J].
Li, Duo ;
Hu, Jie ;
Wang, Changhu ;
Li, Xiangtai ;
She, Qi ;
Zhu, Lei ;
Zhang, Tong ;
Chen, Qifeng .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :12316-12325
[16]  
Li Xiao, 2023, IEEE Trans. Pattern Anal. Mach. Intell.
[17]   Defense against Adversarial Attacks Using High-Level Representation Guided Denoiser [J].
Liao, Fangzhou ;
Liang, Ming ;
Dong, Yinpeng ;
Pang, Tianyu ;
Hu, Xiaolin ;
Zhu, Jun .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :1778-1787
[18]  
Lin Jiadong, 2019, ICLR
[19]  
Liu Y., 2016, ICLR, P1
[20]   Feature Distillation: DNN-Oriented JPEG Compression Against Adversarial Examples [J].
Liu, Zihao ;
Liu, Qi ;
Liu, Tao ;
Xu, Nuo ;
Lin, Xue ;
Wang, Yanzhi ;
Wen, Wujie .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :860-868