Towards Transferable Unrestricted Adversarial Examples with Minimum Changes

被引:3
作者
Liu, Fangcheng [1 ]
Zhang, Chao [1 ]
Zhang, Hongyang [2 ]
机构
[1] Peking Univ, Beijing, Peoples R China
[2] Univ Waterloo, Waterloo, ON, Canada
来源
2023 IEEE CONFERENCE ON SECURE AND TRUSTWORTHY MACHINE LEARNING, SATML | 2023年
基金
加拿大自然科学与工程研究理事会; 国家重点研发计划;
关键词
D O I
10.1109/SaTML54575.2023.00030
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transfer-based adversarial example is one of the most important classes of black-box attacks. However, there is a trade-off between transferability and imperceptibility of the adversarial perturbation. Prior work in this direction often requires a fixed but large l(p)-norm perturbation budget to reach a good transfer success rate, leading to perceptible adversarial perturbations. On the other hand, most of the current unrestricted adversarial attacks that aim to generate semantic-preserving perturbations suffer from weaker transferability to the target model. In this work, we propose a geometry-aware framework to generate transferable adversarial examples with minimum changes. Analogous to model selection in statistical machine learning, we leverage a validation model to select the best perturbation budget for each image under both the l(infinity)- and unrestricted threat models. We propose a principled method for the partition of training and validation models by encouraging intra-group diversity while penalizing extra-group similarity. Extensive experiments verify the effectiveness of our framework on balancing imperceptibility and transferability of the crafted adversarial examples. The methodology is the foundation of our entry to the CVPR'21 Security AI Challenger: Unrestricted Adversarial Attacks on ImageNet, in which we ranked 1st place out of 1,559 teams and surpassed the runner-up submissions by 4.59% and 23.91% in terms of final score and average image quality level, respectively. Code is available at https://github.com/Equationliu/GA-Attack.
引用
收藏
页码:327 / 338
页数:12
相关论文
共 79 条
  • [1] Alaifari R., 2019, ICLR
  • [2] Andriushchenko Maksym, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12368), P484, DOI 10.1007/978-3-030-58592-1_29
  • [3] [Anonymous], 2015, BMVC 2015 P BRIT MAC, DOI 10.5244/c.29.41
  • [4] Athalye A, 2018, PR MACH LEARN RES, V80
  • [5] Brown TB, 2018, Arxiv, DOI arXiv:1809.08352
  • [6] Bai Yutong, 2021, NeurIPS
  • [7] Bhattacharya A., 2020, ICLR, P1, DOI DOI 10.1109/ICCECE48148.2020.9223105
  • [8] Understanding Robustness of Transformers for Image Classification
    Bhojanapalli, Srinadh
    Chakrabarti, Ayan
    Glasner, Daniel
    Li, Daliang
    Unterthiner, Thomas
    Veit, Andreas
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 10211 - 10221
  • [9] Bojarski M, 2016, Arxiv, DOI arXiv:1604.07316
  • [10] Brendel W., 2018, ICLR, P1