Meta-BN Net for few-shot learning

被引:6
作者
Gao, Wei [1 ]
Shao, Mingwen [1 ]
Shu, Jun [2 ]
Zhuang, Xinkai [1 ]
机构
[1] China Univ Petr, Sch Comp Sci, Qingdao 266580, Peoples R China
[2] Xi An Jiao Tong Univ, Sch Math & Stat, Xian 710049, Peoples R China
基金
中国国家自然科学基金;
关键词
meta-learning; few-shot learning; batch normalization;
D O I
10.1007/s11704-021-1237-4
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose a lightweight network with an adaptive batch normalization module, called Meta-BN Net, for few-shot classification. Unlike existing few-shot learning methods, which consist of complex models or algorithms, our approach extends batch normalization, an essential part of current deep neural network training, whose potential has not been fully explored. In particular, a meta-module is introduced to learn to generate more powerful affine transformation parameters, known as sigma and beta, in the batch normalization layer adaptively so that the representation ability of batch normalization can be activated. The experimental results on miniImageNet demonstrate that Meta-BN Net not only outperforms the baseline methods at a large margin but also is competitive with recent state-of-the-art few-shot learning methods. We also conduct experiments on Fewshot-CIFAR100 and CUB datasets, and the results show that our approach is effective to boost the performance of weak baseline networks. We believe our findings can motivate to explore the undiscovered capacity of base components in a neural network as well as more efficient few-shot learning methods.
引用
收藏
页数:8
相关论文
共 54 条
[1]  
Ba JL, 2016, arXiv
[2]  
Bendre N, 2020, arXiv
[3]  
Bengio Y., 1991, IJCNN-91-Seattle: International Joint Conference on Neural Networks (Cat. No.91CH3049-4), DOI 10.1109/IJCNN.1991.155621
[4]  
Bronskill John, 2020, P 37 INT C MACH LEAR, P1153
[5]  
Chen W Y., 2019, P 7 INT C LEARNING R
[6]   CLUSTER SEPARATION MEASURE [J].
DAVIES, DL ;
BOULDIN, DW .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1979, 1 (02) :224-227
[7]  
Dhillon G.S., 2020, ICLR
[8]  
Dumoulin V., 2017, P 5 INT C LEARNING R
[9]  
Finn C, 2017, PR MACH LEARN RES, V70
[10]  
Howard AG, 2017, Arxiv, DOI arXiv:1704.04861