An Efficient End-to-End Channel Level Pruning Method for Deep Neural Networks Compression

被引:0
作者
Zeng, Lei [1 ]
Chen, Shi [1 ]
Zeng, Sen [2 ]
机构
[1] Univ Sci & Techonol China, Dept Comp Sci, Hefei, Anhui, Peoples R China
[2] Dept Gree Elect Appliances, Zhuhai, Guangdong, Peoples R China
来源
PROCEEDINGS OF 2019 IEEE 10TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND SERVICE SCIENCE (ICSESS 2019) | 2019年
关键词
component; DNNS; network pruning; channel pruning; classification;
D O I
10.1109/icsess47205.2019.9040742
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNS) have obtained compelling performance among many visual tasks by a significant increase in the computation and memory consumption, which severely impede their applications on resource-constrained systems like smart mobiles or embedded devices. To solve these problems, recent efforts toward compressing DNNS have received increased focus. In this paper, we proposed an effective end-to-end channel pruning approach to compress DNNS. To this end, firstly, we introduce additional auxiliary classifiers to enhance the discriminative power of shallow and intermediate layers. Secondly, we impose L1-regularization on the scaling factors and shifting factors in batch normalization (BN) layer, and adopt the fast and iterative shrinkage-thresholding algorithm (FISTA) to effectively prune the redundant channels. Finally, by forcing selected factors to zero, we can prune the corresponding unimportant channels safely, thus obtaining a compact model. We empirically reveal the prominent performance of our approach with several state-of-the-art DNNS architectures, including VGGNet, and MobileNet, on different datasets. For instance, on cifar10 dataset, the pruned MobileNet achieves 26.9x reduction in model parameters and 3.9x reduction in computational operations with only 0.04"/0 increase of classification error.
引用
收藏
页码:43 / 46
页数:4
相关论文
共 21 条
  • [1] Beck A., 2009, SIAM
  • [2] Denton Emily L, 2014, ADV NEURAL INFORM PR, P1269
  • [3] Han S., 2016, INT C LEARN REPR ICL
  • [4] Han S., 2015, NEURIPS, DOI DOI 10.5555/2969239.2969366
  • [5] Hassibi Babak, 1993, Advances in Neural Information Processing Systems, P164
  • [6] He Y, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P2234
  • [7] Howard AG, 2017, ARXIV
  • [8] Ioffe S, 2015, 32 INT C MACH LEARN
  • [9] Krizhevsky A., 2009, Tech. Rep. TR-2009, P1
  • [10] ImageNet Classification with Deep Convolutional Neural Networks
    Krizhevsky, Alex
    Sutskever, Ilya
    Hinton, Geoffrey E.
    [J]. COMMUNICATIONS OF THE ACM, 2017, 60 (06) : 84 - 90