The Simple Approach to Multi-label Image Classification Using Transfer Learning

被引:0
作者
Fedorenko, Yuriy S. [1 ]
机构
[1] Bauman Moscow State Tech Univ, Baumanskaya 2 Ya,5, Moscow 105005, Russia
来源
ADVANCES IN NEURAL COMPUTATION, MACHINE LEARNING, AND COGNITIVE RESEARCH III | 2020年 / 856卷
关键词
Image recognition; Transfer learning; Adam; One cycle policy; Weight decay; Amsgrad; Test time augmentation; Confusion matrix;
D O I
10.1007/978-3-030-30425-6_24
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The article deals with the problem of image classification on a relatively small dataset. The training deep convolutional neural net from scratch requires a large amount of data. In many cases, the solution to this problem is to use the pretrained network on another big dataset (e.g. ImageNet) and fine-tune it on available data. In the article, we apply this approach to classify advertising banners images. Initially, we reset the weights of the last layer and change its size to match a number of classes in our dataset. Then we train all network, but the learning rate for the last layer is several times more than for other layers. We use Adam optimization algorithm with some modifications. Firstly, applying weight decay instead of L2 regularization (for Adam they are not same) improves the result. Secondly, the division learning rate on the maximum of gradients squares sum instead of just gradients squares sum makes the training process more stable. Experiments have shown that this approach is appropriate for classifying relatively small datasets. Used metrics and test time augmentation are discussed. Particularly we find that confusion matrix is very useful because it gives an understanding of how to modify the train set to increase model quality.
引用
收藏
页码:207 / 213
页数:7
相关论文
共 9 条
[1]  
[Anonymous], ABS151203385 CORR
[2]  
Ayhan M., 2018, MED IM DEEP LEARN C, P278
[3]  
Gupta A., SUPER CONVERGENCE VE
[4]  
Karpathy A., Convolutional neural networks (CNNs / ConvNets)
[5]  
Loshchilov I., 2019, ARXIV171105101V3CSLG
[6]  
Reddi S., 2018, INT C LEARN REPR ICL, P186
[7]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[8]   Cyclical Learning Rates for Training Neural Networks [J].
Smith, Leslie N. .
2017 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2017), 2017, :464-472
[9]   Rethinking the Inception Architecture for Computer Vision [J].
Szegedy, Christian ;
Vanhoucke, Vincent ;
Ioffe, Sergey ;
Shlens, Jon ;
Wojna, Zbigniew .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2818-2826