Frozen Feature Augmentation for Few-Shot Image Classification

被引:6
作者
Bar, Andreas [1 ,2 ]
Houlsby, Neil [1 ]
Dehghani, Mostafa [1 ]
Kumar, Manoj [1 ]
机构
[1] Google DeepMind, London, England
[2] Tech Univ Carolo Wilhelmina Braunschweig, Braunschweig, Germany
来源
2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2024年
关键词
D O I
10.1109/CVPR52733.2024.01519
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training a linear classifier or lightweight model on top of pretrained vision model outputs, so-called 'frozen features', leads to impressive performance on a number of downstream few-shot tasks. Currently, frozen features are not modified during training. On the other hand, when networks are trained directly on images, data augmentation is a standard recipe that improves performance with no substantial overhead. In this paper, we conduct an extensive pilot study on few-shot image classification that explores applying data augmentations in the frozen feature space, dubbed 'frozen feature augmentation (FroFA)', covering twenty augmentations in total. Our study demonstrates that adopting a deceptively simple pointwise FroFA, such as brightness, can improve few-shot performance consistently across three network architectures, three large pre-training datasets, and eight transfer datasets.
引用
收藏
页码:16046 / 16057
页数:12
相关论文
共 76 条
[1]  
[Anonymous], 2018, P ICML STOCKH SWED
[2]  
[Anonymous], 2019, P ICML LONG BEACH CA
[3]  
[Anonymous], 2019, P CVPR, DOI DOI 10.1109/CVPR.2019.00020
[4]  
[Anonymous], 2018, P NEURIPS MONTR QC C
[5]  
[Anonymous], 2021, P NEURIPS, DOI DOI 10.1109/ICIP42928.2021.9506130
[6]  
[Anonymous], 2023, P ICLR KIG RWAND, DOI DOI 10.1145/3610543.3626179
[7]  
[Anonymous], 2019, P EMNLP WORKSH HONG, DOI DOI 10.1145/3287098.3287099
[8]   Improved Few-Shot Visual Classification [J].
Bateni, Peyman ;
Goyal, Raghav ;
Masrani, Vaden ;
Wood, Frank ;
Sigal, Leonid .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :14481-14490
[9]  
Beattie C., 2016, ARXIV
[10]  
Beyer L., 2022, arXiv