Application of Meta-Heuristic Algorithms for Training Neural Networks and Deep Learning Architectures: A Comprehensive Review

被引:0
作者
Mehrdad Kaveh
Mohammad Saadi Mesgari
机构
[1] K. N. Toosi University of Technology,Department of Geodesy and Geomatics
来源
Neural Processing Letters | 2023年 / 55卷
关键词
Deep learning (DL); Artificial neural networks (ANN); Meta-heuristics (MH); Hyper-parameters optimization; Training; And gradient-based back propagation (BP) learning algorithm;
D O I
暂无
中图分类号
学科分类号
摘要
The learning process and hyper-parameter optimization of artificial neural networks (ANNs) and deep learning (DL) architectures is considered one of the most challenging machine learning problems. Several past studies have used gradient-based back propagation methods to train DL architectures. However, gradient-based methods have major drawbacks such as stucking at local minimums in multi-objective cost functions, expensive execution time due to calculating gradient information with thousands of iterations and needing the cost functions to be continuous. Since training the ANNs and DLs is an NP-hard optimization problem, their structure and parameters optimization using the meta-heuristic (MH) algorithms has been considerably raised. MH algorithms can accurately formulate the optimal estimation of DL components (such as hyper-parameter, weights, number of layers, number of neurons, learning rate, etc.). This paper provides a comprehensive review of the optimization of ANNs and DLs using MH algorithms. In this paper, we have reviewed the latest developments in the use of MH algorithms in the DL and ANN methods, presented their disadvantages and advantages, and pointed out some research directions to fill the gaps between MHs and DL methods. Moreover, it has been explained that the evolutionary hybrid architecture still has limited applicability in the literature. Also, this paper classifies the latest MH algorithms in the literature to demonstrate their effectiveness in DL and ANN training for various applications. Most researchers tend to extend novel hybrid algorithms by combining MHs to optimize the hyper-parameters of DLs and ANNs. The development of hybrid MHs helps improving algorithms performance and capable of solving complex optimization problems. In general, the optimal performance of the MHs should be able to achieve a suitable trade-off between exploration and exploitation features. Hence, this paper tries to summarize various MH algorithms in terms of the convergence trend, exploration, exploitation, and the ability to avoid local minima. The integration of MH with DLs is expected to accelerate the training process in the coming few years. However, relevant publications in this way are still rare.
引用
收藏
页码:4519 / 4622
页数:103
相关论文
共 1660 条
  • [21] Snášel V(2020)Solving the local positioning problem using a four-layer artificial neural network Eng J Geospat Inf Technol 7 21-48
  • [22] Darwish A(1997)The past, present, and future of neural networks for signal processing IEEE Signal Process Mag 14 28-1709
  • [23] Hassanien AE(2011)Nonlinear system identification using memetic differential evolution trained neural networks Neurocomputing 74 1696-2997
  • [24] Das S(2016)Training wavelet neural networks using hybrid particle swarm optimization and gravitational search algorithm for system identification Int J Mechatron Electr Comput Technol 6 2987-114
  • [25] Liu W(1994)Introduction to the special issue on neural networks for speech processing IEEE Trans Speech Audio Process 2 113-100
  • [26] Wang Z(2018)Evolutionary optimization of convolutional neural networks for cancer miRNA biomarkers classification Appl Soft Comput 65 91-934
  • [27] Liu X(2019)Evolutionary generative adversarial networks IEEE Trans Evol Comput 23 921-1314
  • [28] Zeng N(2017)Particle swarm optimization-based automatic parameter selection for deep neural networks and its applications in large-scale and high-dimensional data PLoS ONE 12 e0188746-123
  • [29] Liu Y(2018)Effective long short-term memory with differential evolution algorithm for electricity price prediction Energy 162 1301-191
  • [30] Alsaadi FE(2019)cPSO-CNN: An efficient PSO-based algorithm for fine-tuning hyper-parameters of convolutional neural networks Swarm Evol Comput 49 114-1311