Generalized Gradient Flow Based Saliency for Pruning Deep Convolutional Neural Networks

被引:0
作者
Xinyu Liu
Baopu Li
Zhen Chen
Yixuan Yuan
机构
[1] The Chinese University of Hong Kong,Department of Electronic Engineering
[2] Oracle Cloud Infrastructure (OCI),Centre for Artificial Intelligence and Robotics (CAIR), Hong Kong Institute of Science and Innovation
[3] Chinese Academy of Sciences,undefined
来源
International Journal of Computer Vision | 2023年 / 131卷
关键词
Gradient flow; Model pruning; Network architecture; Normalization; Image classification;
D O I
暂无
中图分类号
学科分类号
摘要
Model filter pruning has shown efficiency in compressing deep convolutional neural networks by removing unimportant filters without sacrificing the performance. However, most existing criteria are empirical, and overlook the relationship between channel saliencies and the non-linear activation functions within the networks. To address these problems, we propose a novel channel pruning method coined gradient flow based saliency (GFBS). Instead of relying on the magnitudes of the entire feature maps, GFBS evaluates the channel saliencies from the gradient flow perspective and only requires the information in normalization and activation layers. Concretely, we first integrate the effects of normalization and ReLU activation layers into convolutional layers based on Taylor expansion. Then, through backpropagation, the derived channel saliency of each layer is indicated by of the first-order Taylor polynomial of the scaling parameter and the signed shifting parameter in the normalization layers. To validate the efficiency and generalization ability of GFBS, we conduct extensive experiments on various tasks, including image classification (CIFAR, ImageNet), image denoising, object detection, and 3D object classification. GFBS could feasibly cooperate with the baseline networks and compress them with only negligible performance drop. Moreover, we extended our method to pruning scratch networks and GFBS is capable to identify subnetworks with comparable performance with the baseline model at an early training stage. Our code has been released at https://github.com/CUHK-AIM-Group/GFBS.
引用
收藏
页码:3121 / 3135
页数:14
相关论文
共 38 条
  • [1] Crawley JN(1985)Cholecystokinin potentiates dopamine-mediated behaviors: Evidence for modulation specific to a site of coexistence Journal of Neuroscience 5 1972-1983
  • [2] Stivers JA(2012)How does the brain solve visual object recognition? Neuron 73 415-434
  • [3] Blumstein LK(2010)The Pascal visual object classes (voc) challenge International Journal of Computer Vision 88 303-338
  • [4] DiCarlo JJ(2021)Jointpruning: Pruning networks along multiple dimensions for efficient point cloud processing IEEE Transactions on Circuits and Systems for Video Technology 4 5947-242
  • [5] Zoccolan D(2009)Deep belief networks Scholarpedia 1 239-12
  • [6] Rust NC(1990)A simple procedure for pruning back-propagation trained neural networks IEEE Transactions on Neural Networks 1 1-58
  • [7] Everingham M(2018)Activations of deep convolutional neural networks are aligned with gamma band activity of human visual cortex Communications Biology 100 53-252
  • [8] Van Gool L(1989)Coexistence of gaba and peptide immunoreactivity in non-pyramidal neurons of the basolateral amygdala Neuroscience Letters 115 211-983
  • [9] Williams CK(2015)Imagenet large scale visual recognition challenge International Journal of Computer Vision 209 976-10,947
  • [10] Guo J(1980)Brain peptides as neurotransmitters Science 33 10,936-1069