An optimal-score-based filter pruning for deep convolutional neural networks

被引:0
作者
Shrutika S. Sawant
J. Bauer
F. X. Erick
Subodh Ingaleshwar
N. Holzer
A. Ramming
E. W. Lang
Th. Götz
机构
[1] Fraunhofer IIS,Department of Internal Medicine 3, Rheumatology & Immunology
[2] Fraunhofer Institute for Integrated Circuits IIS,CIML Group, Biophysics
[3] JSS Academy of Technical Education,undefined
[4] Friedrich-Alexander-University (FAU) Erlangen-Nürnberg and University Hospital Erlangen,undefined
[5] University of Regensburg,undefined
来源
Applied Intelligence | 2022年 / 52卷
关键词
CNN; Deep learning; Filter pruning; Image segmentation; Model compression; Redundancy;
D O I
暂无
中图分类号
学科分类号
摘要
Convolutional Neural Networks (CNN) have achieved excellent performance in the processing of high-resolution images. Most of these networks contain many deep layers in quest of greater segmentation performance. However, over-sized CNN models result in overwhelming memory usage and large inference costs. Earlier studies have revealed that over-sized deep neural models tend to deal with abundant redundant filters that are very similar and provide tiny or no contribution in accelerating the inference of the model. Therefore, we have proposed a novel optimal-score-based filter pruning (OSFP) approach to prune redundant filters according to their relative similarity in feature space. OSFP not only speeds up learning in the network but also eradicates redundant filters leading to improvement in the segmentation performance. We empirically demonstrate on widely used segmentation network models (TernausNet, classical U-Net and VGG16 U-Net) and benchmark datasets (Inria Aerial Image Labeling Dataset and Aerial Imagery for Roof Segmentation (AIRS)) that computation costs (in terms of Float Point Operations (FLOPs) and parameters) are reduced significantly, while maintaining or even improving accuracy.
引用
收藏
页码:17557 / 17579
页数:22
相关论文
共 48 条
  • [1] Ayinde BO(2019)Redundant feature pruning for accelerated inference in deep neural networks Neural Netw 118 148-158
  • [2] Inanc T(2019)Aerial imagery for roof segmentation: a large-scale dataset towards automatic mapping of buildings ISPRS J Photogramm Remote Sens 147 42-55
  • [3] Zurada JM(2020)Deep neural network acceleration based on low-rank Approximated Channel pruning IEEE Trans Circuits Syst I: Regular Papers 67 1232-1244
  • [4] Chen Q(2021)Compressing convolutional neural networks by pruning density peak filters IEEE Access 9 8278-8285
  • [5] Chen Z(2020)Pruning convolutional filters using batch Bridgeout IEEE Access 8 212003-212012
  • [6] Jang Y(2021)Pruning filters with L1-norm and capped L1-norm for CNN compression Appl Intell 51 1152-1160
  • [7] Lee S(1990)Optimal brain damage Adv Neural Inf Proces Syst 2 598-605
  • [8] Kim J(2020)Filter pruning and re-initialization via latent space clustering IEEE Access 8 189587-189597
  • [9] Khan N(2021)Filter pruning-based two-step feature map reconstruction SIViP 15 1555-1563
  • [10] Stavness I(2019)A spectral clustering based filter-level pruning method for convolutional neural networks IEICE Trans Inf Syst E102D 2624-2627