Scalable training of artificial neural networks with adaptive sparse connectivity inspired by network science

被引:331
作者
Mocanu, Decebal Constantin [1 ,2 ]
Mocanu, Elena [2 ,3 ]
Stone, Peter [4 ]
Nguyen, Phuong H. [2 ]
Gibescu, Madeleine [2 ]
Liotta, Antonio [5 ]
机构
[1] Eindhoven Univ Technol, Dept Math & Comp Sci, De Rondom 70, NL-5612 AP Eindhoven, Netherlands
[2] Eindhoven Univ Technol, Dept Elect Engn, De Rondom 70, NL-5612 AP Eindhoven, Netherlands
[3] Eindhoven Univ Technol, Dept Mech Engn, De Rondom 70, NL-5612 AP Eindhoven, Netherlands
[4] Univ Texas Austin, Dept Comp Sci, 2317 Speedway,Stop D9500, Austin, TX 78712 USA
[5] Univ Derby, Data Sci Ctr, Lonsdale House,Quaker Way, Derby DE1 3HD, England
关键词
BRAIN NETWORKS; DEEP; GAME;
D O I
10.1038/s41467-018-04316-3
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Through the success of deep learning in various domains, artificial neural networks are currently among the most used artificial intelligence methods. Taking inspiration from the network properties of biological neural networks (e.g. sparsity, scale-freeness), we argue that (contrary to general practice) artificial neural networks, too, should not have fully-connected layers. Here we propose sparse evolutionary training of artificial neural networks, an algorithm which evolves an initial sparse topology (Erdos-Renyi random graph) of two consecutive layers of neurons into a scale-free topology, during learning. Our method replaces artificial neural networks fully-connected layers with sparse ones before training, reducing quadratically the number of parameters, with no decrease in accuracy. We demonstrate our claims on restricted Boltzmann machines, multi-layer perceptrons, and convolutional neural networks for unsupervised and supervised learning on 15 datasets. Our approach has the potential to enable artificial neural networks to scale up beyond what is currently possible.
引用
收藏
页数:12
相关论文
共 68 条
[1]   Statistical mechanics of complex networks [J].
Albert, R ;
Barabási, AL .
REVIEWS OF MODERN PHYSICS, 2002, 74 (01) :47-97
[2]  
[Anonymous], 2011, JMLR WORKSHOP C P
[3]  
[Anonymous], 2012, REPRESENTATION LEARN
[4]  
[Anonymous], 2017, PREPRINT
[5]  
[Anonymous], 1961, PRINCIPLES NEURODYNA
[6]  
[Anonymous], 2010, J. Mach. Learn. Res.
[7]  
[Anonymous], 2016, COMPUT SCI
[8]  
[Anonymous], 2016, Network Science
[9]  
[Anonymous], 2014, GROWING ADAPTIVE MAC
[10]  
[Anonymous], 1986, Information processing in dynamical systems: Foundations of harmony theory