CFOA: Exploring transferable adversarial examples by content feature optimization with attention guidance

被引:0
|
作者
Liu, Wanping [1 ]
Wang, Baojuan [1 ]
Huang, Dong [2 ]
Luo, Haolan [1 ]
Lu, Ling [1 ]
机构
[1] Chongqing Univ Technol, Coll Comp Sci & Engn, Chongqing 400054, Peoples R China
[2] Guizhou Univ, Key Lab Adv Mfg Technol, Minist Educ, Guiyang 550025, Peoples R China
关键词
Adversarial examples; Transferability; Feature-level perturbations; Parameterized content features; Attention loss;
D O I
10.1016/j.cose.2024.103882
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks are known to be highly susceptible to adversarial examples, created by adding small and carefully-designed perturbations on original images. In black-box scenarios, attacking neural networks is challenging since the information about models is not available. Generating transferable adversarial examples on surrogate models is a viable way. However, current mainstream transfer-based attacks seldomly consider the relationship between transferability and content features of the image. To effectively enhance the attack success rate at the feature level, we newly propose an attack method called Content Feature Optimization Attack (CFOA). This novel CFOA approach operates on white-box models, aiming to generate transferable adversarial examples by targeting the content feature representations of images in the feature space. The method involves the extraction and parameterization of content features, followed by the iterative generation of feature-level perturbations guided by attention loss and adversarial loss. Ultimately, the adversarial examples generated by CFOA exhibit strong transferability on other black-box models. Experimental results show that our method produces adversarial examples with higher success rates in transferability compared to state-of-the-art methods, and importantly, they can effectively evade certain defense mechanisms, presenting a challenge for adversarial defense.
引用
收藏
页数:10
相关论文
共 4 条
  • [1] Rethinking the optimization objective for transferable adversarial examples from a fuzzy perspective
    Yang, Xiangyuan
    Lin, Jie
    Zhang, Hanlin
    Zhao, Peng
    NEURAL NETWORKS, 2025, 184
  • [2] Feature-Based Adversarial Training for Deep Learning Models Resistant to Transferable Adversarial Examples
    Ryu, Gwonsang
    Choi, Daeseon
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (05) : 1039 - 1049
  • [3] Feature Distillation in Deep Attention Network Against Adversarial Examples
    Chen, Xin
    Weng, Jian
    Deng, Xiaoling
    Luo, Weiqi
    Lan, Yubin
    Tian, Qi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (07) : 3691 - 3705
  • [4] Multi-layer Feature Augmentation Based Transferable Adversarial Examples Generation for Speaker Recognition
    Li, Zhuhai
    Zhang, Jie
    Guo, Wu
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT IV, ICIC 2024, 2024, 14865 : 373 - 385