ViP: Unified Certified Detection and Recovery for Patch Attack with Vision Transformers

被引:3
|
作者
Li, Junbo [1 ]
Zhang, Huan [2 ]
Xie, Cihang [1 ]
机构
[1] Univ Calif Santa Cruz, Santa Cruz, CA 95064 USA
[2] Carnegie Mellon Univ, Pittsburgh, PA USA
来源
COMPUTER VISION, ECCV 2022, PT XXV | 2022年 / 13685卷
关键词
Certified defense; Patch attacks; Vision transformer;
D O I
10.1007/978-3-031-19806-9_33
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Patch attack, which introduces a perceptible but localized change to the input image, has gained significant momentum in recent years. In this paper, we present a unified framework to analyze certified patch defense tasks, including both certified detection and certified recovery, leveraging the recently emerged Vision Transformers (ViTs). In addition to the existing patch defense setting where only one patch is considered, we provide the very first study on developing certified detection against the dual patch attack, in which the attacker is allowed to adversarially manipulate pixels in two different regions. By building upon the latest progress in self-supervised ViTs with masked image modeling (i.e., masked autoencoder (MAE)), our method achieves state-of-the-art performance in both certified detection and certified recovery of adversarial patches. Regarding certified detection, we improve the performance by up to similar to 16% on ImageNet without training on a single adversarial patch, and for the first time, can also tackle the more challenging dual patch setting. Our method largely closes the gap between detection-based certified robustness and clean image accuracy. Regarding certified recovery, our approach improves certified accuracy by similar to 2% on ImageNet across all attack sizes, attaining the new state-of-the-art performance.
引用
收藏
页码:573 / 587
页数:15
相关论文
共 34 条
  • [1] IntelPVT: intelligent patch-based pyramid vision transformers for object detection and classification
    Nimma, Divya
    Zhou, Zhaoxian
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (05) : 1767 - 1778
  • [2] IntelPVT: intelligent patch-based pyramid vision transformers for object detection and classification
    Divya Nimma
    Zhaoxian Zhou
    International Journal of Machine Learning and Cybernetics, 2024, 15 : 1767 - 1778
  • [3] Life regression based patch slimming for vision transformers
    Chen, Jiawei
    Chen, Lin
    Yang, Jiang
    Shi, Tianqi
    Cheng, Lechao
    Feng, Zunlei
    Song, Mingli
    NEURAL NETWORKS, 2024, 176
  • [4] TIA: Token Importance Transferable Attack on Vision Transformers
    Fu, Tingchao
    Li, Fanxiao
    Zhang, Jinhong
    Zhu, Liang
    Wang, Yuanyu
    Zhou, Wei
    INFORMATION SECURITY AND CRYPTOLOGY, INSCRYPT 2023, PT II, 2024, 14527 : 91 - 107
  • [5] PIPformers: Patch based inpainting with vision transformers for generalize paintings
    Lee, Jeyoung
    Kang, Hochul
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2024, 35 (03)
  • [6] Towards Optimal Patch Size in Vision Transformers for Tumor Segmentation
    Mojtahedi, Ramtin
    Hamghalam, Mohammad
    Do, Richard K. G.
    Simpson, Amber L.
    MULTISCALE MULTIMODAL MEDICAL IMAGING, MMMI 2022, 2022, 13594 : 110 - 120
  • [7] Presentation attack detection based on two-stream vision transformers with self-attention fusion
    Peng, Fei
    Meng, Shao-hua
    Long, Min
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 85
  • [8] Colonoscopy Landmark Detection Using Vision Transformers
    Tamhane, Aniruddha
    Mida, Tse'ela
    Posner, Erez
    Bouhnik, Moshe
    IMAGING SYSTEMS FOR GI ENDOSCOPY, AND GRAPHS IN BIOMEDICAL IMAGE ANALYSIS, ISGIE 2022, 2022, 13754 : 24 - 34
  • [9] Improved robustness of vision transformers via prelayernorm in patch emb e dding
    Kim, Bum Jun
    Choi, Hyeyeon
    Jang, Hyeonah
    Lee, Dong Gu
    Jeong, Wonseok
    Kim, Sang Woo
    PATTERN RECOGNITION, 2023, 141
  • [10] Vision transformers are active learners for image copy detection
    Tan, Zhentao
    Wang, Wenhao
    Shan, Caifeng
    NEUROCOMPUTING, 2024, 587