Convergence Analysis of PSO for Hyper-Parameter Selection in Deep Neural Networks

被引:5
作者
Nalepa, Jakub [1 ,2 ]
Lorenzo, Pablo Ribalta [1 ]
机构
[1] Future Proc, Gliwice, Poland
[2] Silesian Tech Univ, Gliwice, Poland
来源
ADVANCES ON P2P, PARALLEL, GRID, CLOUD AND INTERNET COMPUTING (3PGCIC-2017) | 2018年 / 13卷
关键词
Convergence analysis; PSO; Hyper; parameter selection; DNNs;
D O I
10.1007/978-3-319-69835-9_27
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks (DNNs) have gained enormous research attention since they consistently outperform other state-of-the-art methods in a plethora of machine learning tasks. However, their performance strongly depends on the DNN hyper-parameters which are commonly tuned by experienced practitioners. Recently, we introduced Particle Swarm Optimization (PSO) and parallel PSO techniques to automate this process. In this work, we theoretically and experimentally investigate the convergence capabilities of these algorithms. The experiments were performed for several DNN architectures (both gradually augmented and hand-crafted by a human) using two challenging multi-class benchmark datasets-MNIST and CIFAR-10.
引用
收藏
页码:284 / 295
页数:12
相关论文
共 15 条
[1]  
[Anonymous], 2007, IEEE INT C ICML
[2]  
[Anonymous], 2012, P 25 INT C NEURIPS
[3]  
[Anonymous], ADV NEURAL INF PROCE
[4]  
Bergstra J, 2012, J MACH LEARN RES, V13, P281
[5]  
Ciresan D, 2012, PROC CVPR IEEE, P3642, DOI 10.1109/CVPR.2012.6248110
[6]  
David E., 2014, GECCO 2014 COMP PUBL, DOI [DOI 10.1145/2598394.2602287, 10.1145/2598394.2602287]
[7]  
Graham B., 2014, Fractional Max-Pooling
[8]  
Ilievski Ilija, 2016, CIRED WORKSHOP 2016, P1
[9]  
King DB, 2015, ACS SYM SER, V1214, P1
[10]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324