Class-Separation Preserving Pruning for Deep Neural Networks

被引:0
作者
Preet I. [1 ,2 ]
Boydell O. [1 ]
John D. [3 ]
机构
[1] University College Dublin, CeADAR - Ireland's Centre for Applied AI, Dublin
[2] Eaton Corporation Plc., Dublin
[3] University College Dublin, School of Electrical and Electronics Engineering, Dublin
来源
IEEE Transactions on Artificial Intelligence | 2024年 / 5卷 / 01期
关键词
Class-separation score (CSS); deep neural networks (DNNs); pruning; structured pruning;
D O I
10.1109/TAI.2022.3228511
中图分类号
学科分类号
摘要
Neural network pruning has been deemed essential in the deployment of deep neural networks on resource-constrained edge devices, greatly reducing the number of network parameters without drastically compromising accuracy. A class of techniques proposed in the literature assigns an importance score to each parameter and prunes those of the least importance. However, most of these methods are based on generalized estimations of the importance of each parameter, ignoring the context of the specific task at hand. In this article, we propose a task specific pruning approach, CSPrune, which is based on how efficiently a neuron or a convolutional filter is able to separate classes. Our axiomatic approach assigns an importance score based on how separable different classes are in the output activations or feature maps, preserving the separation of classes which avoids the reduction in classification accuracy. Additionally, most pruning algorithms prune individual connections or weights leading to a sparse network without taking into account whether the hardware the network is deployed on can take advantage of that sparsity or not. CSPrune prunes whole neurons or filters which results in a more structured pruned network whose sparsity can be more efficiently utilized by the hardware. We evaluate our pruning method against various benchmark datasets, both small and large, and network architectures and show that our approach outperforms comparable pruning techniques. © 2020 IEEE.
引用
收藏
页码:290 / 299
页数:9
相关论文
共 50 条
[41]   Structured pruning of neural networks for constraints learning [J].
Cacciola, Matteo ;
Frangioni, Antonio ;
Lodi, Andrea .
OPERATIONS RESEARCH LETTERS, 2024, 57
[42]   Automated Pruning of Neural Networks for Mobile Applications [J].
Glinserer, Andreas ;
Lechner, Martin ;
Wendt, Alexander .
2021 IEEE 19TH INTERNATIONAL CONFERENCE ON INDUSTRIAL INFORMATICS (INDIN), 2021,
[43]   Pruning neural networks for inductive conformal prediction [J].
Zhao, Xindi ;
Bellotti, Anthony .
CONFORMAL AND PROBABILISTIC PREDICTION WITH APPLICATIONS, VOL 179, 2022, 179
[44]   Pruning Coherent Integrated Photonic Neural Networks [J].
Banerjee, Sanmitra ;
Nikdast, Mahdi ;
Pasricha, Sudeep ;
Chakrabarty, Krishnendu .
IEEE JOURNAL OF SELECTED TOPICS IN QUANTUM ELECTRONICS, 2023, 29 (02)
[45]   PRUNING ARTIFICIAL NEURAL NETWORKS USING NEURAL COMPLEXITY MEASURES [J].
Jorgensen, Thomas D. ;
Haynes, Barry P. ;
Norlund, Charlotte C. F. .
INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2008, 18 (05) :389-403
[46]   Joint Pruning and Channel-Wise Mixed-Precision Quantization for Efficient Deep Neural Networks [J].
Motetti, Beatrice Alessandra ;
Risso, Matteo ;
Burrello, Alessio ;
Macii, Enrico ;
Poncino, Massimo ;
Pagliari, Daniele Jahier .
IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (11) :2619-2633
[47]   Adaptive training and pruning for neural networks: algorithms and application [J].
Chen, S ;
Chang, SJ ;
Yuan, JH ;
Zhang, YX ;
Wong, KW .
ACTA PHYSICA SINICA, 2001, 50 (04) :674-681
[48]   Self-organizing neural networks by construction and pruning [J].
Lee, JS ;
Lee, H ;
Kim, JY ;
Nam, D ;
Park, CH .
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2004, E87D (11) :2489-2498
[49]   Pruning feature maps for efficient convolutional neural networks [J].
Guo, Xiao-ting ;
Xie, Xin-shu ;
Lang, Xun .
OPTIK, 2023, 281
[50]   Sensitivity-Informed Provable Pruning of Neural Networks [J].
Baykal, Cenk ;
Liebenwein, Lucas ;
Gilitschenski, Igor ;
Feldman, Dan ;
Rus, Daniela .
SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2022, 4 (01) :26-45