Push & Pull: Transferable Adversarial Examples With Attentive Attack

被引:29
|
作者
Gao, Lianli [1 ,2 ,3 ]
Huang, Zijie [2 ,3 ]
Song, Jingkuan [1 ]
Yang, Yang [2 ,3 ]
Shen, Heng Tao [2 ,3 ]
机构
[1] Univ Elect Sci & Technol China, Inst Neurol, Sichuan Prov Peoples Hosp, Chengdu 611731, Peoples R China
[2] Univ Elect Sci & Technol China, Future Media Ctr, Chengdu 611731, Peoples R China
[3] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Feature extraction; Computational modeling; Task analysis; Predictive models; Neural networks; Iterative methods; Image classification; adversarial attack; transferability; targeted attack;
D O I
10.1109/TMM.2021.3079723
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Targeted attack aims to mislead the classification model to a specific class, and it can be further divided into black-box and white-box targeted attack depending on whether the classification model is known. A growing number of approaches rely on disrupting the image representations to craft adversarial examples. However, this type of methods often suffer from either low white-box targeted attack success rate or poor black-box targeted attack transferability. To address these problems, we propose a Transferable Attentive Attack (TAA) method which adds perturbation to clean images based on the attended regions and features. This is motivated by one important observation that deep-learning based classification models (or even shallow-learning based models like SIFT) make the prediction mainly based on the informative and discriminative regions of an image. Specifically, the corresponding features of the informative regions are firstly extracted, and the anchor image's features are iteratively "pushed" away from the source class and simultaneously "pulled" closer to the target class along with attacking. Moreover, we introduce a new strategy that the attack selects the centroids of source and target class cluster as the input of triplet loss to achieve high transferability. Experimental results demonstrate that our method improves the transferability of adversarial example, while maintaining higher success rate for white-box targeted attacks compared with the state-of-the-arts. In particular, TAA attacks on image-representation based task like VQA also result in a significant performance drop in terms of accuracy.
引用
收藏
页码:2329 / 2338
页数:10
相关论文
共 50 条
  • [31] AGS: Affordable and Generalizable Substitute Training for Transferable Adversarial Attack
    Wang, Ruikui
    Guo, Yuanfang
    Wang, Yunhong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 5553 - 5562
  • [32] AN ENHANCED TRANSFERABLE ADVERSARIAL ATTACK OF SCALE-INVARIANT METHODS
    Lin, Zhi
    Peng, Anjie
    Wei, Rong
    Yu, Wenxin
    Zeng, Hui
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 3788 - 3792
  • [33] A Transferable Adversarial Belief Attack With Salient Region Perturbation Restriction
    Zhang, Shihui
    Zuo, Dongxu
    Yang, Yongliang
    Zhang, Xiaowei
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4296 - 4306
  • [34] Towards the transferable audio adversarial attack via ensemble methods
    Guo, Feng
    Sun, Zheng
    Chen, Yuxuan
    Ju, Lei
    CYBERSECURITY, 2023, 6 (01)
  • [35] Black-box Bayesian adversarial attack with transferable priors
    Zhang, Shudong
    Gao, Haichang
    Shu, Chao
    Cao, Xiwen
    Zhou, Yunyi
    He, Jianping
    MACHINE LEARNING, 2024, 113 (04) : 1511 - 1528
  • [36] Toward Transferable Attack via Adversarial Diffusion in Face Recognition
    Hu, Cong
    Li, Yuanbo
    Feng, Zhenhua
    Wu, Xiaojun
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 5506 - 5519
  • [37] Towards the transferable audio adversarial attack via ensemble methods
    Feng Guo
    Zheng Sun
    Yuxuan Chen
    Lei Ju
    Cybersecurity, 6
  • [38] Towards Transferable Adversarial Attack Against Deep Face Recognition
    Zhong, Yaoyao
    Deng, Weihong
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 : 1452 - 1466
  • [39] Greedy Attack and Gumbel Attack: Generating Adversarial Examples for Discrete Data
    Yang, Puyudi
    Chen, Jianbo
    Hsieh, Cho-Jui
    Wang, Jane-Ling
    Jordan, Michael, I
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [40] A method for filtering the attack pairs of adversarial examples based on attack distance
    Liu H.
    Fang Y.
    Wen W.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2022, 48 (02): : 339 - 347