Structured Pruning of Deep Convolutional Neural Networks

被引:426
作者
Anwar, Sajid [1 ]
Hwang, Kyuyeon [1 ]
Sung, Wonyong [1 ]
机构
[1] Seoul Natl Univ, Dept Elect Engn & Comp Sci, 1 Gwanak Ro, Seoul 08826, South Korea
基金
新加坡国家研究基金会;
关键词
Deep convolutional neural networks; structured pruning; feature map pruning; intra-kernel strided sparsity; PARTICLE FILTERS;
D O I
10.1145/3005348
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Real-time application of deep learning algorithms is often hindered by high computational complexity and frequent memory accesses. Network pruning is a promising technique to solve this problem. However, pruning usually results in irregular network connections that not only demand extra representation efforts but also do not fit well on parallel computation. We introduce structured sparsity at various scales for convolutional neural networks: feature map-wise, kernel-wise, and intra-kernel strided sparsity. This structured sparsity is very advantageous for direct computational resource savings on embedded computers, in parallel computing environments, and in hardware-based systems. To decide the importance of network connections and paths, the proposed method uses a particle filtering approach. The importance weight of each particle is assigned by assessing the misclassification rate with a corresponding connectivity pattern. The pruned network is retrained to compensate for the losses due to pruning. While implementing convolutions as matrix products, we particularly show that intra-kernel strided sparsity with a simple constraint can significantly reduce the size of the kernel and feature map tensors. The proposed work shows that when pruning granularities are applied in combination, we can prune the CIFAR-10 network by more than 70% with less than a 1% loss in accuracy.
引用
收藏
页数:18
相关论文
共 33 条
  • [1] [Anonymous], ARXIV150602515
  • [2] [Anonymous], 2014, ARXIV NEURAL EVOLUTI
  • [3] [Anonymous], 2015, ARXIV151000149
  • [4] [Anonymous], 2011, Proceedings of the Conference on Neural Information Processing Systems Workshops
  • [5] [Anonymous], 2014, arXiv
  • [6] [Anonymous], 2013, Fast training of convolutional networks through FFTs
  • [7] A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking
    Arulampalam, MS
    Maskell, S
    Gordon, N
    Clapp, T
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2002, 50 (02) : 174 - 188
  • [8] Improved particle filter for nonlinear problems
    Carpenter, J
    Clifford, P
    Fearnhead, P
    [J]. IEE PROCEEDINGS-RADAR SONAR AND NAVIGATION, 1999, 146 (01) : 2 - 7
  • [9] An iterative pruning algorithm for feedforward neural networks
    Castellano, G
    Fanelli, AM
    Pelillo, M
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1997, 8 (03): : 519 - 531
  • [10] Chellapilla K., 2006, P ICFHR