SAPENet: Self-Attention based Prototype Enhancement Network for Few-shot Learning

被引:44
作者
Huang, Xilang [1 ]
Choi, Seon Han [2 ,3 ]
机构
[1] Pukyong Natl Univ, Dept Artificial Intelligent Convergence, Pusan 48513, South Korea
[2] Ewha Womans Univ, Dept Elect & Elect Engn, Seoul 03760, South Korea
[3] Ewha Womans Univ, Grad Program Smart Factory, Seoul 03760, South Korea
基金
新加坡国家研究基金会;
关键词
Few -shot learning; Multi -head self -attention mechanism; Image classification; k -Nearest neighbor;
D O I
10.1016/j.patcog.2022.109170
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot learning considers the problem of learning unseen categories given only a few labeled samples. As one of the most popular few-shot learning approaches, Prototypical Networks have received considerable attention owing to their simplicity and efficiency. However, a class prototype is typically obtained by averaging a few labeled samples belonging to the same class, which treats the samples as equally important and is thus prone to learning redundant features. Herein, we propose a self-attention based prototype enhancement network (SAPENet) to obtain a more representative prototype for each class. SAPENet utilizes multi-head self-attention mechanisms to selectively augment discriminative features in each sample feature map, and generates channel attention maps between intra-class sample features to attentively retain informative channel features for that class. The augmented feature maps and attention maps are finally fused to obtain representative class prototypes. Thereafter, a local descriptor-based metric module is employed to fully exploit the channel information of the prototypes by searching k similar local descriptors of the prototype for each local descriptor in the unlabeled samples for classification. We performed experiments on multiple benchmark datasets: miniImageNet, tieredImageNet, and CUB-200-2011. The experimental results on these datasets show that SAPENet achieves a considerable improvement compared to Prototypical Networks and also outperforms related state-of-the-art methods.(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页数:11
相关论文
共 40 条
[1]  
Afrasiyabi A., 2022, PROC IEEECVF C COMPU, P9014
[2]  
Chen W. Y., 2019, P INT C LEARNING REP
[3]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[4]  
Finn C, 2017, PR MACH LEARN RES, V70
[5]   Dual Attention Network for Scene Segmentation [J].
Fu, Jun ;
Liu, Jing ;
Tian, Haijie ;
Li, Yong ;
Bao, Yongjun ;
Fang, Zhiwei ;
Lu, Hanqing .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3141-3149
[6]   Collect and Select: Semantic Alignment Metric Learning for Few-Shot Learning [J].
Hao, Fusheng ;
He, Fengxiang ;
Cheng, Jun ;
Wang, Lei ;
Cao, Jianzhong ;
Tao, Dacheng .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :8459-8468
[7]  
He J., 2022, ACM T MULTIM COMPUT
[8]  
Hou RB, 2019, ADV NEUR IN, V32
[9]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/TPAMI.2019.2913372, 10.1109/CVPR.2018.00745]
[10]   Local descriptor-based multi-prototype network for few-shot Learning [J].
Huang, Hongwei ;
Wu, Zhangkai ;
Li, Wenbin ;
Huo, Jing ;
Gao, Yang .
PATTERN RECOGNITION, 2021, 116