Efficient hyperparameter optimization through model-based reinforcement learning

被引:57
作者
Wu, Jia [1 ]
Chen, SenPeng [1 ]
Liu, XiYuan [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu, Peoples R China
关键词
Hyperparameter optimization; Machine learning; Reinforcement learning;
D O I
10.1016/j.neucom.2020.06.064
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hyperparameter tuning is critical for the performance of machine learning algorithms. However, a noticeable limitation is the high computational cost of algorithm evaluation for complex models or for large datasets, which makes the tuning process highly inefficient. In this paper, we propose a novel model-based method for efficient hyperparameter optimization. Firstly, we frame this optimization process as a reinforcement learning problem and then employ an agent to tune hyperparameters sequentially. In addition, a model that learns how to evaluate an algorithm is used to speed up the training. However, model inaccuracy is further exacerbated by long-term use, resulting in collapse performance. We propose a novel method for controlling the model use by measuring the impact of the model on the policy and limiting it to a proper range. Thus, the horizon of the model use can be dynamically adjusted. We apply the proposed method to tune the hyperparameters of the extreme gradient boosting and convolutional neural networks on 101 tasks. The experimental results verify that the proposed method achieves the highest accuracy on 86.1% of the tasks, compared with other state-of-the-art methods and the average ranking of runtime is significant lower than all methods by using the predictive model. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页码:381 / 393
页数:13
相关论文
共 46 条
[1]  
[Anonymous], 2018, STUD CONFL DEV PEACE
[2]  
[Anonymous], 2019, IEEE INTERNET THINGS, DOI DOI 10.1109/JIOT.2019.2905743
[3]  
Asfour T., INT C LEARN REPR 18
[4]  
Bergstra J, 2012, J MACH LEARN RES, V13, P281
[5]  
Brian M., 2017, P 32 AAAI C ART INT, P378
[6]  
Cai H, 2018, AAAI CONF ARTIF INTE, P2787
[7]  
Chang X., ARXIV170511159
[8]   Ensemble-based deep reinforcement learning for chatbots [J].
Cuayahuitl, Heriberto ;
Lee, Donghyeon ;
Ryu, Seonghan ;
Cho, Yongjin ;
Choi, Sungja ;
Indurthi, Satish ;
Yu, Seunghak ;
Choi, Hyungtak ;
Hwang, Inchul ;
Kim, Jihie .
NEUROCOMPUTING, 2019, 366 :118-130
[9]  
Deisenroth M.P., 2011, INT C INT C MACH LEA
[10]  
Demsar J, 2006, J MACH LEARN RES, V7, P1