Meta-BN Net for few-shot learning

被引:0
作者
Wei GAO [1 ]
Mingwen SHAO [1 ]
Jun SHU [2 ]
Xinkai ZHUANG [1 ]
机构
[1] School of Computer Science,China University of Petroleum
[2] School of Mathematics and Statistics,Xi'an Jiaotong University
关键词
meta-learning; few-shot learning; batch normalization;
D O I
暂无
中图分类号
TP183 [人工神经网络与计算]; TP391.41 [];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ; 080203 ;
摘要
In this paper,we propose a lightweight network with an adaptive batch normalization module,called Meta-BN Net,for few-shot classification.Unlike existing few-shot learning methods,which consist of complex models or algorithms,our approach extends batch normalization,an essential part of current deep neural network training,whose potential has not been fully explored.In particular,a meta-module is introduced to learn to generate more powerful affine transformation parameters,known as γ and β,in the batch normalization layer adaptively so that the representation ability of batch normalization can be activated.The experimental results on miniImageNet demonstrate that Meta-BN Net not only outperforms the baseline methods at a large margin but also is competitive with recent state-of-the-art few-shot learning methods.We also conduct experiments on Fewshot-CIFAR100 and CUB datasets,and the results show that our approach is effective to boost the performance of weak baseline networks.We believe our findings can motivate to explore the undiscovered capacity of base components in a neural network as well as more efficient few-shot learning methods.
引用
收藏
页码:76 / 83
页数:8
相关论文
共 9 条
[1]  
DeepEMD: Few-shot image classification with differentiable earth mover's distance and structured classifiers.[J].Zhang C.;Cai Y.;Lin G.;Shen C..Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition.2020,
[2]  
Instance Normalization: The Missing Ingredient for Fast Stylization..[J].Dmitry Ulyanov;Andrea Vedaldi;Victor S. Lempitsky.CoRR.2016,
[3]  
One-shot Learning with Memory-Augmented Neural Networks..[J].Adam Santoro;Sergey Bartunov;Matthew Botvinick;Daan Wierstra;Timothy P. Lillicrap.CoRR.2016,
[4]  
Layer Normalization..[J].Lei Jimmy Ba;Ryan Kiros;Geoffrey E. Hinton.CoRR.2016,
[5]   ImageNet Large Scale Visual Recognition Challenge [J].
Russakovsky, Olga ;
Deng, Jia ;
Su, Hao ;
Krause, Jonathan ;
Satheesh, Sanjeev ;
Ma, Sean ;
Huang, Zhiheng ;
Karpathy, Andrej ;
Khosla, Aditya ;
Bernstein, Michael ;
Berg, Alexander C. ;
Fei-Fei, Li .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 115 (03) :211-252
[6]   Human-level concept learning through probabilistic program induction [J].
Lake, Brenden M. ;
Salakhutdinov, Ruslan ;
Tenenbaum, Joshua B. .
SCIENCE, 2015, 350 (6266) :1332-1338
[7]  
Neural Turing Machines..[J].Alex Graves;Greg Wayne;Ivo Danihelka.CoRR.2014,
[8]   One-shot learning of object categories [J].
Li, FF ;
Fergus, R ;
Perona, P .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2006, 28 (04) :594-611
[9]   CLUSTER SEPARATION MEASURE [J].
DAVIES, DL ;
BOULDIN, DW .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1979, 1 (02) :224-227