Retraining-free methods for fast on-the-fly pruning of convolutional neural networks

被引:11
作者
Ashouri, Amir H. [1 ]
Abdelrahman, Tarek S. [2 ,3 ]
Dos Remedios, Alwyn [4 ]
机构
[1] Univ Toronto, ECE Dept, Toronto, ON, Canada
[2] Univ Toronto, Elect & Comp Engn, Toronto, ON, Canada
[3] Univ Toronto, Comp Sci, Toronto, ON, Canada
[4] Qualcomm Inc, Markham, ON, Canada
关键词
Deep learning; Convolutional neural networks; Sparsity; Pruning;
D O I
10.1016/j.neucom.2019.08.063
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We explore retraining-free pruning of CNNs. We propose and evaluate three model-independent methods for sparsification of model weights. Our methods are magnitude-based, efficient, and can be applied on-the-fly during model load time, which is necessary in some deployment contexts. We evaluate the effectiveness of these methods in introducing sparsity with minimal loss of inference accuracy using five state-of-the-art pretrained CNNs. The evaluation shows that the methods reduce the number of weights by up to 73% (i.e., compression factor of 3.7 x) without incurring more than 5% loss in Top-5 accuracy. These results also hold for quantized versions of the CNNs. We develop a classifier to determine which of the three methods is most suited for a given model. Finally, we employ additional, but impractical in our deployment context, fine-tuning and show that it gains only 8% in sparsity. This indicates that our on-the-fly methods capture much of the sparsity than can be attained without retraining, yet remain efficient and straight-forward to use. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:56 / 69
页数:14
相关论文
共 50 条
  • [1] Acceleration-aware, Retraining-free Evolutionary Pruning for Automated Fitment of Deep Learning Models on Edge Devices
    Dutta, Jeet
    Dey, Swarnava
    Mukherjee, Arijit
    Pal, Arpan
    SECOND INTERNATIONAL CONFERENCE ON AIML SYSTEMS 2022, 2022,
  • [2] Pruning convolutional neural networks via filter similarity analysis
    Lili Geng
    Baoning Niu
    Machine Learning, 2022, 111 : 3161 - 3180
  • [3] Pruning convolutional neural networks via filter similarity analysis
    Geng, Lili
    Niu, Baoning
    MACHINE LEARNING, 2022, 111 (09) : 3161 - 3180
  • [4] Leveraging Structured Pruning of Convolutional Neural Networks
    Tessier, Hugo
    Gripon, Vincent
    Leonardon, Mathieu
    Arzel, Matthieu
    Bertrand, David
    Hannagan, Thomas
    2022 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS), 2022, : 174 - 179
  • [5] Structured Pruning for Deep Convolutional Neural Networks: A Survey
    He, Yang
    Xiao, Lingao
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (05) : 2900 - 2919
  • [6] Metaheuristics for pruning convolutional neural networks: A comparative study
    Palakonda, Vikas
    Tursunboev, Jamshid
    Kang, Jae-Mo
    Moon, Sunghwan
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 268
  • [8] On-the-Fly Syntax Highlighting using Neural Networks
    Palma, Marco Edoardo
    Salza, Pasquale
    Gall, Harald C.
    PROCEEDINGS OF THE 30TH ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2022, 2022, : 269 - 280
  • [9] Fast Convolutional Neural Networks in Low Density FPGAs Using Zero-Skipping and Weight Pruning
    Vestias, Mario P.
    Duarte, Rui Policarpo
    de Sousa, Jose T.
    Neto, Horacio C.
    ELECTRONICS, 2019, 8 (11)
  • [10] Blending Pruning Criteria for Convolutional Neural Networks
    He, Wei
    Huang, Zhongzhan
    Liang, Mingfu
    Liang, Senwei
    Yang, Haizhao
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 3 - 15