A Cyclical Learning Rate Method in Deep Learning Training

被引:11
作者
Li, Jiaqi [1 ]
Yang, Xiaodong [1 ]
机构
[1] ZhejiangGongshang Univ, Sch Informat & Elect Engn, Hangzhou, Peoples R China
来源
PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON COMPUTER, INFORMATION AND TELECOMMUNICATION SYSTEMS (CITS) | 2020年
关键词
Deep learning; Learning rate; Accuracy; Deep neural networks;
D O I
10.1109/cits49457.2020.9232482
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The learning rate is an important hyperparameter for training deep neural networks. The traditional learning rate method has the problems of instability of accuracy. Aiming at these problems, we proposed a new learning rate method with different cyclical changes in each training cycle instead of a fixed value. It achieves higher accuracy in less iterations and faster convergence. Through the experiment on CIFAR-10 and CIFAR-100 datasets based on VGG network and RESNET network, the final results show that the proposed method has better results on stability and accuracy than cyclical learning rate method.
引用
收藏
页码:140 / 144
页数:5
相关论文
共 21 条
[1]  
Allen-Zhu Z., 2019, INT C MACH LEARN
[2]  
[Anonymous], 2012, LECT 6 5 RMSPROP NEU
[3]  
[Anonymous], 2014, LOSS SURFACE MULTILA
[4]  
Chen Y.-A., ARXIV PREPRINT ARXIV
[5]  
Chung J., 2015, ARXIV PREPRINT ARXIV
[6]  
Dinh L, 2017, PR MACH LEARN RES, V70
[7]  
Duchi J, 2011, J MACH LEARN RES, V12, P2121
[8]  
George AP, 2006, MACH LEARN, V65, P167, DOI 10.1007/S10994-006-8365-9
[9]  
Gulcehre C, 2014, EPRINT ARXIV
[10]  
Jastrzbski S, 2017, THREE FACTORS INFLUE