Channel-augmented joint transformation for transferable adversarial attacks

被引:0
作者
Desheng Zheng
Wuping Ke
Xiaoyu Li
Shibin Zhang
Guangqiang Yin
Weizhong Qian
Yong Zhou
Fan Min
Shan Yang
机构
[1] Southwest Petroleum University,School of Computer Science
[2] University of Electronic Science and Technology of China,School of Information and Software Engineering
[3] Silk Road Talent Building,Institute of Electronics and Information Industry Technology of Kash
[4] Advanced Cryptography and System Security Key Laboratory of Sichuan Province,Department of Chemistry, Physics and Atmospheric Science
[5] Chengdu University of Information Technology,undefined
[6] Jackson State University,undefined
来源
Applied Intelligence | 2024年 / 54卷
关键词
Adversarial examples; Neural networks; Computer vision; AI security;
D O I
暂无
中图分类号
学科分类号
摘要
Deep neural networks (DNNs) are vulnerable to adversarial examples that fool the models with tiny perturbations. Although adversarial attacks have achieved incredible attack success rates in the white-box setting, most existing adversaries often exhibit weak transferability in the black-box setting, especially for models with defense mechanisms. In this work, we reveal the cross-model channel redundancy and channel invariance of DNNs and thus propose two channel-augmented methods to improve the transferability of adversarial examples, namely, the channel transformation (CT) method and the channel-invariant Patch (CIP) method. Specifically, channel transformation shuffles and rewrites channels to enhance cross-model feature redundancy in convolution, and channel-invariant patches distinctly weaken different channels to achieve loss-preserving transformation. We compute the aggregated gradients of the transformed dataset to create adversaries with higher transferability. In addition, the two proposed methods can be naturally combined with each other and with almost all other gradient-based methods to further improve performance. Empirical results on the ImageNet dataset demonstrate that our attack methods exhibit higher transferability and achieve higher attack success rates than state-of-the-art gradient-based attacks. Specifically, our attack improves the average attack success rate from 86.9% to 91.0% on normally trained models and from 44.6% to 68.3% on adversarially trained models.
引用
收藏
页码:428 / 442
页数:14
相关论文
共 50 条
[1]  
Touvron H(2022)Resmlp: Feedforward networks for image classi cation with data-efficient training IEEE Trans Pattern Anal Mach Intell 45 5314-5321
[2]  
Bojanowski P(2023)Pv-rcnn++: Point-voxel feature set abstraction with local vector rep- resentation for 3d object detection Int J Comput Vis 131 531-551
[3]  
Caron M(2022)Scale-aware automatic augmentations for object detection with dynamic training IEEE Trans Pattern Anal Mach Intell 45 2367-2383
[4]  
Cord M(2023)Semi-supervised and long-tailed object detection with cascadematch Int J Comput Vision 131 987-1001
[5]  
El-Nouby A(2023)Minimally distorted structured adversarial attacks Int J Comput Vision 131 160-176
[6]  
Grave E(2022)Random and adversarial bit error robustness: Energy-efficient and secure dnn accelerators IEEE Trans Pattern Anal Mach Intell 45 3632-3647
[7]  
Izacard G(2022)Graddiv: Adversarial robustness of randomized neural networks via gradient diversity regularization IEEE Trans Pattern Anal Mach Intell 45 2645-2651
[8]  
Joulin A(2023)No-reference qquality index of tone-mapped images based on authenticity, preservation, and scene expressiveness Signal Process 203 6434-6445
[9]  
Synnaeve G(2021)Slimconv: Reducing channel redundancy in convolutional neural networks by features recombining IEEE Trans Image Process 30 undefined-undefined
[10]  
Verbeek J(undefined)undefined undefined undefined undefined-undefined