Optimizing Relevance Maps of Vision Transformers Improves Robustness

被引:0
|
作者
Chefer, Hila [1 ]
Schwartz, Idan [1 ]
Wolf, Lior [1 ]
机构
[1] Tel Aviv Univ, Sch Comp Sci, Tel Aviv, Israel
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
基金
欧洲研究理事会;
关键词
DECISIONS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It has been observed that visual classification models often rely mostly on spurious cues such as the image background, which hurts their robustness to distribution changes. To alleviate this shortcoming, we propose to monitor the model's relevancy signal and direct the model to base its prediction on the foreground object. This is done as a finetuning step, involving relatively few samples consisting of pairs of images and their associated foreground masks. Specifically, we encourage the model's relevancy map (i) to assign lower relevance to background regions, (ii) to consider as much information as possible from the foreground, and (iii) we encourage the decisions to have high confidence. When applied to Vision Transformer (ViT) models, a marked improvement in robustness to domain-shifts is observed. Moreover, the foreground masks can be obtained automatically, from a self-supervised variant of the ViT model itself; therefore no additional supervision is required. Our code is available at: https://github.com/hila-chefer/RobustViT.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Understanding The Robustness in Vision Transformers
    Zhou, Daquan
    Yu, Zhiding
    Xie, Enze
    Xiao, Chaowei
    Anandkumar, Anima
    Feng, Jiashi
    Alvarez, Jose M.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [2] On the Robustness of Vision Transformers to Adversarial Examples
    Mahmood, Kaleel
    Mahmood, Rigel
    van Dijk, Marten
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7818 - 7827
  • [3] Certified Patch Robustness via Smoothed Vision Transformers
    Salman, Hadi
    Jain, Saachi
    Wong, Eric
    Madry, Aleksander
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15116 - 15126
  • [4] Harnessing Edge Information for Improved Robustness in Vision Transformers
    Li, Yanxi
    Du, Chengbin
    Xu, Chang
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 4, 2024, : 3252 - 3260
  • [5] Optimizing Mobile Vision Transformers for Land Cover Classification
    Rozario, Papia F.
    Gadgil, Ravi
    Lee, Junsu
    Gomes, Rahul
    Keller, Paige
    Liu, Yiheng
    Sipos, Gabriel
    Mcdonnell, Grace
    Impola, Westin
    Rudolph, Joseph
    APPLIED SCIENCES-BASEL, 2024, 14 (13):
  • [6] Adversarial Robustness of Vision Transformers Versus Convolutional Neural Networks
    Ali, Kazim
    Bhatti, Muhammad Shahid
    Saeed, Atif
    Athar, Atifa
    Al Ghamdi, Mohammed A.
    Almotiri, Sultan H.
    Akram, Samina
    IEEE ACCESS, 2024, 12 : 105281 - 105293
  • [7] Understanding Adversarial Robustness of Vision Transformers via Cauchy Problem
    Wang, Zheng
    Ruan, Wenjie
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT III, 2023, 13715 : 562 - 577
  • [8] On the robustness of vision transformers for in-flight monocular depth estimation
    Simone Ercolino
    Alessio Devoto
    Luca Monorchio
    Matteo Santini
    Silvio Mazzaro
    Simone Scardapane
    Industrial Artificial Intelligence, 1 (1):
  • [9] Trade-off between Robustness and Accuracy of Vision Transformers
    Li, Yanxi
    Xu, Chang
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 7558 - 7568
  • [10] Vision transformers in domain adaptation and domain generalization: a study of robustness
    Alijani, Shadi
    Fayyad, Jamil
    Najjaran, Homayoun
    Neural Computing and Applications, 2024, 36 (29) : 17979 - 18007