SIM: an improved few-shot image classification model with multi-task learning

被引:3
作者
Guo, Jin [1 ,2 ]
Li, Wengen [1 ]
Guan, Jihong [1 ]
Gao, Hang [2 ,3 ]
Liu, Baobo [2 ]
Gong, Lili [4 ]
机构
[1] Tongji Univ, Dept Comp Sci & Technol, Shanghai, Peoples R China
[2] Tongji Univ, Project Management Off, China Natl Sci Seafloor Observ, Shanghai, Peoples R China
[3] Tongji Univ, Sch Ocean & Earth Sci, State Key Lab Marine Geol, Shanghai, Peoples R China
[4] Shanghai Publishing & Printing Coll, Off Acad Res, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
few-shot; self-supervised learning; image classification; contrastive learning; multi-task learning;
D O I
10.1117/1.JEI.31.3.033044
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Few-shot learning (FSL) has been widely used for image classification when the training samples are limited and can effectively avoid overfitting. Most FSL methods are based on metric learning and tend to learn features that are conducive to classifying known classes of images. However, these methods cannot capture the semantic features that are important for classifying new classes of images. To address this issue, we proposed an improved few-shot image classification model based on multi-task learning termed SIM. It combines the self-supervised image representation learning task with the supervised image classification task, thus utilizing the complementarily of these two tasks. SIM has two stages, i.e., the pre-training stage and the meta-training stage. In pre-training stage, we learned the representation of training images via supervised learning and self-supervised learning. Then, in meta-training stage, we trained a linear classifier based on the learned representation. According to the experiments on four data sets, including three natural image data sets and one marine plankton image data set, SIM outperformed existing methods and achieved quite good performance in complex application scenarios. (C) 2022 SPIE and IS&T
引用
收藏
页数:13
相关论文
共 31 条
[1]   Deep Clustering for Unsupervised Learning of Visual Features [J].
Caron, Mathilde ;
Bojanowski, Piotr ;
Joulin, Armand ;
Douze, Matthijs .
COMPUTER VISION - ECCV 2018, PT XIV, 2018, 11218 :139-156
[2]  
Chen T, 2020, PR MACH LEARN RES, V119
[3]   Multi-task Self-Supervised Visual Learning [J].
Doersch, Carl ;
Zisserman, Andrew .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2070-2079
[4]   Unsupervised Visual Representation Learning by Context Prediction [J].
Doersch, Carl ;
Gupta, Abhinav ;
Efros, Alexei A. .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1422-1430
[5]  
Finn C, 2017, PR MACH LEARN RES, V70
[6]  
Gidaris S., 2018, INT C LEARN REPR ICL
[7]  
Grill Jean-Bastien, 2020, P INT C MACH LEARN
[8]  
He K., 2017, P IEEE INT C COMPUTE, P2980, DOI [DOI 10.1109/ICCV.2017.322, 10.1109/ICCV.2017.322]
[9]   UberNet: Training a Universal Convolutional Neural Network for Low-, Mid-, and High-Level Vision using Diverse Datasets and Limited Memory [J].
Kokkinos, Iasonas .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :5454-5463
[10]   Human-level concept learning through probabilistic program induction [J].
Lake, Brenden M. ;
Salakhutdinov, Ruslan ;
Tenenbaum, Joshua B. .
SCIENCE, 2015, 350 (6266) :1332-1338