Few-Shot Learning With Enhancements to Data Augmentation and Feature Extraction

被引:10
作者
Zhang, Yourun [1 ,2 ]
Gong, Maoguo [3 ,4 ]
Li, Jianzhao [4 ]
Feng, Kaiyuan [4 ]
Zhang, Mingyang [4 ]
机构
[1] Xidian Univ, Hangzhou Inst Technol, Hangzhou 311200, Peoples R China
[2] Xidian Univ, Key Lab Collaborat Intelligence Syst, Minist Educ, Xian 710071, Peoples R China
[3] Xidian Univ, Key Lab Collaborat Intelligence Syst, Minist Educ, Xian 010028, Peoples R China
[4] Xidian Univ, Key Lab Collaborat Intelligence Syst, Minist Educ, Xian 710071, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Training; Data models; Testing; Manifolds; Data augmentation; Task analysis; feature extraction; few-shot learning; image classification; ALIGNMENT;
D O I
10.1109/TNNLS.2024.3400592
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The few-shot image classification task is to enable a model to identify novel classes by using only a few labeled samples as references. In general, the more knowledge a model has, the more robust it is when facing novel situations. Although directly introducing large amounts of new training data to acquire more knowledge is an attractive solution, it violates the purpose of few-shot learning with respect to reducing dependence on big data. Another viable option is to enable the model to accumulate knowledge more effectively from existing data, i.e., improve the utilization of existing data. In this article, we propose a new data augmentation method called self-mixup (SM) to assemble different augmented instances of the same image, which facilitates the model to more effectively accumulate knowledge from limited training data. In addition to the utilization of data, few-shot learning faces another challenge related to feature extraction. Specifically, existing metric-based few-shot classification methods rely on comparing the extracted features of the novel classes, but the widely adopted downsampling structures in various networks can lead to feature degradation due to the violation of the sampling theorem, and the degraded features are not conducive to robust classification. To alleviate this problem, we propose a calibration-adaptive downsampling (CADS) that calibrates and utilizes the characteristics of different features, which can facilitate robust feature extraction and benefit classification. By improving data utilization and feature extraction, our method shows superior performance on four widely adopted few-shot classification datasets.
引用
收藏
页码:6655 / 6668
页数:14
相关论文
共 70 条
[1]  
Bendou Y., 2022, ARXIV
[2]  
Bertinetto L., 2019, INT C LEARNING REPRE, P1
[3]   Hierarchical Graph Neural Networks for Few-Shot Learning [J].
Chen, Cen ;
Li, Kenli ;
Wei, Wei ;
Zhou, Joey Tianyi ;
Zeng, Zeng .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (01) :240-252
[4]  
Chen W., 2019, 7 INT C LEARN REPR I
[5]   Meta-Baseline: Exploring Simple Meta-Learning for Few-Shot Learning [J].
Chen, Yinbo ;
Liu, Zhuang ;
Xu, Huijuan ;
Darrell, Trevor ;
Wang, Xiaolong .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :9042-9051
[6]   Image Deformation Meta-Networks for One-Shot Learning [J].
Chen, Zitian ;
Fu, Yanwei ;
Wang, Yu-Xiong ;
Ma, Lin ;
Liu, Wei ;
Hebert, Martial .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :8672-8681
[7]   SPNet: Siamese-Prototype Network for Few-Shot Remote Sensing Image Scene Classification [J].
Cheng, Gong ;
Cai, Liming ;
Lang, Chunbo ;
Yao, Xiwen ;
Chen, Jinyong ;
Guo, Lei ;
Han, Junwei .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
[8]   Learning to Capture the Query Distribution for Few-Shot Learning [J].
Chi, Ziqiu ;
Wang, Zhe ;
Yang, Mengping ;
Li, Dongdong ;
Du, Wenli .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (07) :4163-4173
[9]  
Faramarzi Mojtaba, 2020, ARXIV
[10]  
Finn C, 2017, PR MACH LEARN RES, V70