Robust Few-Shot Learning Without Using Any Adversarial Samples

被引:2
作者
Nayak, Gaurav Kumar [1 ]
Rawal, Ruchit [2 ]
Khatri, Inder [3 ]
Chakraborty, Anirban [4 ]
机构
[1] Univ Cent Florida, Ctr Res Comp Vis, Orlando, FL 32816 USA
[2] Max Planck Inst Software Syst, D-66123 Saarbrucken, Germany
[3] NYU, Tandon Sch Engn, New York, NY 10012 USA
[4] Indian Inst Sci, Dept Computat & Data Sci, Bengaluru 560012, India
关键词
Robustness; Training; Metalearning; Computational modeling; Optimization; Task analysis; Standards; Adversarial defense; adversarial robustness; few-shot learning; Fourier transform; self distillation;
D O I
10.1109/TNNLS.2023.3336996
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The high cost of acquiring and annotating samples has made the "few-shot" learning problem of prime importance. Existing works mainly focus on improving performance on clean data and overlook robustness concerns on the data perturbed with adversarial noise. Recently, a few efforts have been made to combine the few-shot problem with the robustness objective using sophisticated meta-learning techniques. These methods rely on the generation of adversarial samples in every episode of training, which further adds to the computational burden. To avoid such time-consuming and complicated procedures, we propose a simple but effective alternative that does not require any adversarial samples. Inspired by the cognitive decision-making process in humans, we enforce high-level feature matching between the base class data and their corresponding low-frequency samples in the pretraining stage via self distillation. The model is then fine-tuned on the samples of novel classes where we additionally improve the discriminability of low-frequency query set features via cosine similarity. On a one-shot setting of the CIFAR-FS dataset, our method yields a massive improvement of 60.55% and 62.05% in adversarial accuracy on the projected gradient descent (PGD) and state-of-the-art auto attack, respectively, with a minor drop in clean accuracy compared to the baseline. Moreover, our method only takes 1.69x of the standard training time while being approximate to 5x faster than thestate-of-the-art adversarial meta-learning methods. The code is available at https://github.com/vcl-iisc/robust-few-shot-learning.
引用
收藏
页码:2080 / 2090
页数:11
相关论文
共 43 条
[1]   Towards Achieving Adversarial Robustness by Enforcing Feature Consistency Across Bit Planes [J].
Addepalli, Sravanti ;
Vivek, B. S. ;
Baburaj, Arya ;
Sriramanan, Gaurang ;
Babu, R. Venkatesh .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :1017-1026
[2]  
Alayrac J.-B., 2019, P ADV NEUR INF PROC, V32, P110
[3]  
Allen-Zhu Z., 2023, P INT C LEARN REPR I, P1
[4]  
Bengio Y., 2012, P ICML WORKSH UNS TR, V7, P19
[5]  
Bertinetto L., 2019, P 7 INT C LEARN REPR
[6]  
Carmon Y., 2019, P ADV NEURALINF PROC, P1
[7]  
Chandavarkar B.R., 2020, 2020 11 INT C COMP C, P1, DOI [10.1109/ICCCNT49239.2020.9225520, DOI 10.1109/ICCCNT49239.2020.9225520]
[8]   Mimic and Fool: A Task-Agnostic Adversarial Attack [J].
Chaturvedi, Akshay ;
Garain, Utpal .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (04) :1801-1808
[9]   SMGEA: A New Ensemble Adversarial Attack Powered by Long-Term Gradient Memories [J].
Che, Zhaohui ;
Borji, Ali ;
Zhai, Guangtao ;
Ling, Suiyi ;
Li, Jing ;
Min, Xiongkuo ;
Guo, Guodong ;
Le Callet, Patrick .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (03) :1051-1065
[10]   Improving Adversarial Robustness via Guided Complement Entropy [J].
Chen, Hao-Yun ;
Liang, Jhao-Hong ;
Chang, Shih-Chieh ;
Pan, Jia-Yu ;
Chen, Yu-Ting ;
Wei, Wei ;
Juan, Da-Cheng .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :4880-4888