Application of Monte Carlo stochastic optimization (MOST) to deep learning

被引:7
作者
Inage, Sin-ichi [1 ,2 ]
Hebishima, Hana [1 ]
机构
[1] Fukuoka Univ, Dept Mech Engn, Fluid Engn Lab, Fukuoka 8140180, Japan
[2] Fukuoka Univ, Dept Mech Engn, 8-19-1,Nanakuma,Jonan Ku, Fukuoka 8140180, Japan
关键词
Optimization algorithms; Neural network; Deep learning; Monte Carlo method; Genetic algorithm; Adam;
D O I
10.1016/j.matcom.2022.03.013
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In this paper, we propose a new optimization method based on the Monte Carlo method. The proposed method is applied to several benchmark problems, and the result of applying it to the optimization of neural network is reported. Deep machine learning using neural networks is one of the important keywords to promote innovation in today's advanced information society. Therefore, research on large-scale, high-speed, and high-precision algorithms has been actively conducted. The author has developed an optimization method in which the search region for multivariate parameters constituting the objective function is divided into two regions for each parameter, the integral values of each divided region are numerically calculated by the Monte Carlo method, the magnitude of each integral value is compared, and the optimum point is judged to be in a small region. The proposed method was applied to 50 variable benchmark functions (Schwefel and Ackley Functions), and was compared with the results of genetic algorithm (GA) which was a representative of existing optimization methods. As a result, it was confirmed that the proposed method is faster and more accurate than GA. In addition, the proposed method is applied to machine learning by neural networks, specifically XOR gate circuits and IRS classification problems, and verified. The neural network optimized by MOST reproduced teacher data and test data faster and more accurately than conventional Adam and genetic algorithms. (c) 2022 International Association for Mathematics and Computers in Simulation (IMACS). Published by Elsevier B.V. All rights reserved.
引用
收藏
页码:257 / 271
页数:15
相关论文
共 19 条
[1]  
[Anonymous], US
[2]  
[Anonymous], ADADELTA: An Adaptive Learning Rate Method
[3]  
Bahar Parnia, 2017, Prague Bulletin of Mathematical Linguistics, P13, DOI 10.1515/pralin-2017-0005
[4]  
Bottou L, 2012, OPTIMIZATION FOR MACHINE LEARNING, P351
[5]  
Dozat T., 2016, INCORPORATING NESTER
[6]  
Duchi J, 2011, J MACH LEARN RES, V12, P2121
[7]  
Frenkel Daan, 2004, INTRO MONTE CARLO ME
[8]  
Gouldberg D., 1989, GENERIC ALGORITHM SE
[9]  
Hayashi Hiroaki, 2018, EVE GRADIENT BASED O
[10]  
Henderson Peter, 2018, EUROPEAN WORKSHOP RE, P2018