Hyperparameter Optimization: Comparing Genetic Algorithm against Grid Search and Bayesian Optimization

被引:191
作者
Alibrahim, Hussain [1 ]
Ludwig, Simone A. [1 ]
机构
[1] North Dakota State Univ, Dept Comp Sci, Fargo, ND 58105 USA
来源
2021 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC 2021) | 2021年
关键词
Hyperparmeter optimization; Grid Search; Bayesian; Genetic Algorithm;
D O I
10.1109/CEC45853.2021.9504761
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The performance of machine learning algorithms are affected by several factors, some of these factors are related to data quantity, quality, or its features. Another element is the choice of an appropriate algorithm to solve the problem and one major influence is the parameter configuration based on the problem specification. Parameters in machine learning can be classified in two types: (1) model parameters that are internal, configurable, and its value can be estimated from data such as weights of a deep neural network; and (2) hyperparameters, which are external and its values can not be estimated from data such as the learning rate for the training of a neural network. Hyperparameter values may be specified by a practitioner or using a heuristic, or parameter values obtained from other problems can be used etc., however, the best values of these parameters are identified when the algorithm has the highest accuracy, and these could be achieved by tuning the parameters. The main goal of this paper is to conduct a comparison study between different algorithms that are used in the optimization process in order to find the best hyperparameter values for the neural network. The algorithms applied are grid search algorithm, bayesian algorithm, and genetic algorithm. Different evaluation measures are used to conduct this comparison such as accuracy and running time.
引用
收藏
页码:1551 / 1559
页数:9
相关论文
共 21 条
  • [1] Bergstra J, 2012, J MACH LEARN RES, V13, P281
  • [2] Resampling Methods for Meta-Model Validation with Recommendations for Evolutionary Computation
    Bischl, B.
    Mersmann, O.
    Trautmann, H.
    Weihs, C.
    [J]. EVOLUTIONARY COMPUTATION, 2012, 20 (02) : 249 - 275
  • [3] Buluswar S. D., 1994, Image Understanding Workshop. Proceedings, P1619, DOI 10.1117/12.188926
  • [4] Chollet F., 2015, KERAS 20 COMPUTER SO
  • [5] DIFRANCESCOMARI.C, 2018, PUBLICATION INFORM S
  • [6] Joachims T, 1998, P 10 EUR C MACH LEAR
  • [7] Kegl B., 2014, NEURAL INFORM PROCES
  • [8] A Study on L2-Loss (Squared Hinge-Loss) Multiclass SVM
    Lee, Ching-Pei
    Lin, Chih-Jen
    [J]. NEURAL COMPUTATION, 2013, 25 (05) : 1302 - 1323
  • [9] INDUCTIVE TEXT CLASSIFICATION FOR MEDICAL APPLICATIONS
    LEHNERT, W
    SODERLAND, S
    ARONOW, D
    FENG, FF
    SHMUELI, A
    [J]. JOURNAL OF EXPERIMENTAL & THEORETICAL ARTIFICIAL INTELLIGENCE, 1995, 7 (01) : 49 - 80
  • [10] Liu D, 2020, SNDCNN SELF NORMALIZ