Large Scale Evolution of Convolutional Neural Networks Using Volunteer Computing

被引:42
作者
Desell, Travis [1 ]
机构
[1] Univ North Dakota, 3950 Campus Rd, Grand Forks, ND 58201 USA
来源
PROCEEDINGS OF THE 2017 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION (GECCO'17 COMPANION) | 2017年
基金
美国国家科学基金会;
关键词
Neuroevolution; Convolutional Neural Networks; Image Classification; Distributed Evolutionary Algorithms;
D O I
10.1145/3067695.3076002
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work presents a new algorithm called evolutionary exploration of augmenting convolutional topologies (EXACT), which is capable of evolving the structure of convolutional neural networks (CNNs). EXACT is in part modeled after the neuroevolution of augmenting topologies (NEAT) algorithm, with notable exceptions to allow it to scale to large scale distributed computing environments and evolve networks with convolutional filters. In addition to multi-threaded and MPI versions, EXACT has been implemented as part of a BOINC volunteer computing project, allowing large scale evolution. During a period of two months, over 4,500 volunteered computers on the Citizen Science Grid trained over 120,000 CNNs and evolved networks reaching 98.32% test data accuracy on the MNIST handwritten digits dataset. Oese results are even stronger as the backpropagation strategy used to train the CNNs was fairly rudimentary (ReLU units, L2 regularization and Nesterov momentum) and these were initial test runs done without refinement of the backpropagation hyperparameters. Further, the EXACT evolutionary strategy is independent of the method used to train the CNNs, so they could be further improved by advanced techniques like elastic distortions, pretraining and dropout. The evolved networks are also quite interesting, showing "organic" structures and significant differences from standard human designed architectures.
引用
收藏
页码:127 / 128
页数:2
相关论文
共 10 条
[1]  
He K., 2016, P IEEE C COMPUTER VI, P770, DOI DOI 10.1109/CVPR.2016.90
[2]  
Krizhevsky A., 2009, LEARNING MULTIPLE LA
[3]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[4]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324
[5]  
Miikkulainen R, 2019, ARTIFICIAL INTELLIGENCE IN THE AGE OF NEURAL NETWORKS AND BRAIN COMPUTING, P293, DOI 10.1016/B978-0-12-815480-9.00015-3
[6]  
Real E., 2017, 34 INT C MACHINE LEA, P2902
[7]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[8]   Evolving neural networks through augmenting topologies [J].
Stanley, KO ;
Miikkulainen, R .
EVOLUTIONARY COMPUTATION, 2002, 10 (02) :99-127
[9]  
Szegedy C, 2014, Arxiv, DOI [arXiv:1312.6199, DOI 10.1109/CVPR.2015.7298594]
[10]   80 million tiny images: A large data set for nonparametric object and scene recognition [J].
Torralba, Antonio ;
Fergus, Rob ;
Freeman, William T. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2008, 30 (11) :1958-1970