Efficient Hyperparameter Optimization for Convolution Neural Networks in Deep Learning: A Distributed Particle Swarm Optimization Approach

被引:55
作者
Guo, Yu [1 ]
Li, Jian-Yu [1 ]
Zhan, Zhi-Hui [1 ]
机构
[1] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510006, Peoples R China
关键词
Convolution neural network (CNN); deep learning; distributed particle swarm optimization algorithm (DPSO); hyperparameter; particle swarm optimization (PSO); ALGORITHM;
D O I
10.1080/01969722.2020.1827797
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolution neural network (CNN) is a kind of powerful and efficient deep learning approach that has obtained great success in many real-world applications. However, due to its complex network structure, the intertwining of hyperparameters, and the time-consuming procedure for network training, finding an efficient network configuration for CNN is a challenging yet tough work. To efficiently solve the hyperparameters setting problem, this paper proposes a distributed particle swarm optimization (DPSO) approach, which can optimize the hyperparameters to find high-performing CNNs. Compared to tedious, historical-experience-based, and personal-preference-based manual designs, the proposed DPSO approach can evolve the hyperparameters automatically and globally to obtain promising CNNs, which provides a new idea and approach for finding the global optimal hyperparameter combination. Moreover, by cooperating with the distributed computing techniques, the DPSO approach can have a considerable speedup when compared with the traditional particle swarm optimization (PSO) algorithm. Extensive experiments on widely-used image classification benchmarks have verified that the proposed DPSO approach can effectively find the CNN model with promising performance, and at the same time, has greatly reduced the computational time when compared with traditional PSO.
引用
收藏
页码:36 / 57
页数:22
相关论文
共 43 条
  • [1] Genetic Programming With a New Representation to Automatically Learn Features and Evolve Ensembles for Image Classification
    Bi, Ying
    Xue, Bing
    Zhang, Mengjie
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (04) : 1769 - 1783
  • [2] PCANet: A Simple Deep Learning Baseline for Image Classification?
    Chan, Tsung-Han
    Jia, Kui
    Gao, Shenghua
    Lu, Jiwen
    Zeng, Zinan
    Ma, Yi
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2015, 24 (12) : 5017 - 5032
  • [3] Distributed Individuals for Multiple Peaks: A Novel Differential Evolution for Multimodal Optimization Problems
    Chen, Zong-Gan
    Zhan, Zhi-Hui
    Wang, Hua
    Zhang, Jun
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2020, 24 (04) : 708 - 719
  • [4] Clanuwat T., 2018, DEEP LEARNING CLASSI
  • [5] SUPPORT-VECTOR NETWORKS
    CORTES, C
    VAPNIK, V
    [J]. MACHINE LEARNING, 1995, 20 (03) : 273 - 297
  • [6] NEAREST NEIGHBOR PATTERN CLASSIFICATION
    COVER, TM
    HART, PE
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 1967, 13 (01) : 21 - +
  • [7] Solution Modeling Using Postfix Genetic Programming
    Dabhi, Vipul K.
    Chaudhary, Sanjay
    [J]. CYBERNETICS AND SYSTEMS, 2015, 46 (08) : 605 - 640
  • [8] Dorigo M., 1997, IEEE Transactions on Evolutionary Computation, V1, P53, DOI 10.1109/4235.585892
  • [9] Particle swarm optimization of deep neural networks architectures for image classification
    Fernandes Junior, Francisco Erivaldo
    Yen, Gary G.
    [J]. SWARM AND EVOLUTIONARY COMPUTATION, 2019, 49 : 62 - 74
  • [10] Dendritic Neuron Model With Effective Learning Algorithms for Classification, Approximation, and Prediction
    Gao, Shangce
    Zhou, MengChu
    Wang, Yirui
    Cheng, Jiujun
    Yachi, Hanaki
    Wang, Jiahai
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (02) : 601 - 614