NeST: A Neural Network Synthesis Tool Based on a Grow-and-Prune Paradigm

被引:134
作者
Dai, Xiaoliang [1 ]
Yin, Hongxu [1 ]
Jha, Niraj K. [1 ]
机构
[1] Princeton Univ, Dept Elect Engn, Princeton, NJ 08544 USA
基金
美国国家科学基金会;
关键词
Architecture synthesis; grow-and-prune paradigm; network parameters; neural network;
D O I
10.1109/TC.2019.2914438
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks (DNNs) have begun to have a pervasive impact on various applications of machine learning. However, the problem of finding an optimal DNN architecture for large applications is challenging. Common approaches go for deeper and larger DNN architectures but may incur substantial redundancy. To address these problems, we introduce a network growth algorithm that complements network pruning to learn both weights and compact DNN architectures during training. We propose a DNN synthesis tool (NeST) that combines both methods to automate the generation of compact and accurate DNNs. NeSTstarts with a randomly initialized sparse network called the seed architecture. It iteratively tunes the architecture with gradient-based growth and magnitude-based pruning of neurons and connections. Our experimental results show that NeST yields accurate, yet very compact DNNs, with a wide range of seed architecture selection. For the LeNet-300-100 (LeNet-5) architecture, we reduce network parameters by 70.2 x (74.3 x) and floating-point operations (FLOPs) by 79.4 x (43.7 x). For the AlexNet, VGG-16, and ResNet-50 architectures, we reduce network parameters (FLOPs) by 15.7 x (4.6 x), 33.2 x (8.9 x), and 4.1 x (2.1 x) respectively. NeST's grow-and-prune paradigm delivers significant additional parameter and FLOPs reduction relative to pruning-only methods.
引用
收藏
页码:1487 / 1497
页数:11
相关论文
共 40 条
[1]  
Abadi M., 2015, TENSORFLOW LARGE SCA, DOI DOI 10.48550/ARXIV.1603.04467
[2]  
[Anonymous], 2015, ARXIV PREPRINT ARXIV
[3]  
[Anonymous], P NIPS WORKSH AUT
[4]  
[Anonymous], 2015, P 3 INT C LEARN REPR
[5]  
[Anonymous], 2015, P COMP SCI SWANS UK
[6]  
[Anonymous], ADV NEURAL INFORM PR
[7]  
[Anonymous], 2014, arXiv
[8]  
[Anonymous], 2015, PROCIEEE CONFCOMPUT, DOI DOI 10.1109/CVPR.2015.7298594
[9]  
[Anonymous], 2017, P INT C LEARN REPR
[10]  
Anwar S, 2016, ARXIV PREPRINT ARXIV