Hyperparameter optimization of deep neural network using univariate dynamic encoding algorithm for searches

被引:85
作者
Yoo, YoungJun [1 ]
机构
[1] Pohang Univ Sci & Technol POSTECH, Dept Elect Engn, San 31, Pohang 790784, Gyungbuk, South Korea
关键词
Hyperparameter optimization; Gradient-free optimization; Deep neural network; Convolution neural network; Autoencoder;
D O I
10.1016/j.knosys.2019.04.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes a method to find the hyperparameter tuning for a deep neural network by using a univariate dynamic encoding algorithm for searches. Optimizing hyperparameters for such a neural network is difficult because the neural network that has several parameters to configure: furthermore, the training speed for such a network is slow. The proposed method was tested for two neural network models: an autoencoder and a convolution neural network with the Modified National Institute of Standards and Technology (MNIST) dataset. To optimize hyperparameters with the proposed method, the cost functions were selected as the average of the difference between the decoded value and the original image for the autoencoder, and the inverse of the evaluation accuracy for the convolution neural network. The hyperparameters were optimized using the proposed method with fast convergence speed and few computational resources, and the results were compared with those of the other considered optimization algorithms (namely, simulated annealing, genetic algorithm, and particle swarm algorithm) to show the effectiveness of the proposed methodology. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:74 / 83
页数:10
相关论文
共 30 条
[1]   GLOBAL OPTIMIZATION AND STOCHASTIC DIFFERENTIAL-EQUATIONS [J].
ALUFFIPENTINI, F ;
PARISI, V ;
ZIRILLI, F .
JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 1985, 47 (01) :1-16
[2]  
[Anonymous], 1993, Advances in Neural Information Processing Systems
[3]  
[Anonymous], 2015, MANAG SCI LETT, DOI DOI 10.5267/J.MSL.2015.4.002
[4]  
[Anonymous], 1989, Global optimization
[5]   Gradient-based optimization of hyperparameters [J].
Bengio, Y .
NEURAL COMPUTATION, 2000, 12 (08) :1889-1900
[6]  
Bergstra J., 2013, P 30 INT C MACH LEAR, P115
[7]  
Bergstra J., 2011, Advances in Neural Information Processing Systems (NIPS), V24, P1, DOI 10.5555/2986459.2986743
[8]  
Bergstra J, 2012, J MACH LEARN RES, V13, P281
[9]   Selective maintenance scheduling under stochastic maintenance quality with multiple maintenance actions [J].
Duan, Chaoqun ;
Deng, Chao ;
Gharaei, Abolfazl ;
Wu, Jun ;
Wang, Bingran .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2018, 56 (23) :7160-7178
[10]   TABOO SEARCH - AN APPROACH TO THE MULTIPLE MINIMA PROBLEM [J].
CVIJOVIC, D ;
KLINOWSKI, J .
SCIENCE, 1995, 267 (5198) :664-666