An Automatically Layer-Wise Searching Strategy for Channel Pruning Based on Task-Driven Sparsity Optimization

被引:27
作者
Feng, Kai-Yuan [1 ]
Fei, Xia [1 ]
Gong, Maoguo [1 ]
Qin, A. K. [2 ]
Li, Hao [1 ]
Wu, Yue [3 ]
机构
[1] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[2] Swinburne Univ Technol, Dept Comp Technol, Hawthorn, Vic 3122, Australia
[3] Xidian Univ, Sch Comp Sci & Technol, Xian 710071, Peoples R China
基金
中国国家自然科学基金; 澳大利亚研究理事会;
关键词
Task analysis; Knowledge engineering; Training; Cost function; Convolutional neural networks; Computational modeling; Tensors; Deep neural networks; channel pruning; knowledge distillation; compression; NEURAL-NETWORKS;
D O I
10.1109/TCSVT.2022.3156588
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep convolutional neural networks (CNNs) have achieved tremendous successes but tend to suffer from high computation costs mainly due to heavy over-parameterization, resulting in the difficulty of directly applying them to the ever-growing application demands based on low-end edge devices with strong power restriction and real-time inference requirement. Recently, there has much research attention devoted to compressing the network via pruning to address this issue. Most of the existing methods rely on some hand-designed pruning rules, which suffer from several limitations. Firstly, manually designed rules are only applicable to limited application scenarios, which can hardly generalize well in a broader scope. And these rules are typically designed based on human experience and via trial and error, and thus highly subjective. Then, channels of different layers in a network may have diverse distributions, which means the same pruning rule is not appropriate for each layer. To address these limitations, we propose a novel channel pruning scheme, in which the task-irrelevant channels are removed in a task-driven manner. Specifically, an adaptively differentiable search module is proposed to find the best pruning rule automatically for different layers in CNNs under sparsity constraints. Besides, we employed knowledge distillation to alleviate the excessive performance loss. Once the training process is finished, a compact network will be obtained by removing channels based on layer-wise pruning rules. We have evaluated the proposed method on some well-known benchmark datasets including CIFAR, MNIST, and ImageNet in comparison to several state-of-the-art pruning methods. Experimental results demonstrate the superiority of our method over the compared ones in terms of both parameters and FLOPs reduction.
引用
收藏
页码:5790 / 5802
页数:13
相关论文
共 64 条
[41]   Learning Efficient Convolutional Networks through Network Slimming [J].
Liu, Zhuang ;
Li, Jianguo ;
Shen, Zhiqiang ;
Huang, Gao ;
Yan, Shoumeng ;
Zhang, Changshui .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2755-2763
[42]  
Long J, 2015, PROC CVPR IEEE, P3431, DOI 10.1109/CVPR.2015.7298965
[43]  
Luo JH, 2019, Arxiv, DOI arXiv:1805.08941
[44]   ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression [J].
Luo, Jian-Hao ;
Wu, Jianxin ;
Lin, Weiyao .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :5068-5076
[45]  
Luo P, 2016, AAAI CONF ARTIF INTE, P3560
[46]   Image and Video Compression With Neural Networks: A Review [J].
Ma, Siwei ;
Zhang, Xinfeng ;
Jia, Chuanmin ;
Zhao, Zhenghui ;
Wang, Shiqi ;
Wang, Shanshe .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (06) :1683-1698
[47]  
Mei J, 2019, PROC INT C LEARN REP, P1
[48]  
Mirzadeh SI, 2020, AAAI CONF ARTIF INTE, V34, P5191
[49]   XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks [J].
Rastegari, Mohammad ;
Ordonez, Vicente ;
Redmon, Joseph ;
Farhadi, Ali .
COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 :525-542
[50]   ImageNet Large Scale Visual Recognition Challenge [J].
Russakovsky, Olga ;
Deng, Jia ;
Su, Hao ;
Krause, Jonathan ;
Satheesh, Sanjeev ;
Ma, Sean ;
Huang, Zhiheng ;
Karpathy, Andrej ;
Khosla, Aditya ;
Bernstein, Michael ;
Berg, Alexander C. ;
Fei-Fei, Li .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 115 (03) :211-252