FILTER PRUNING BASED ON LOCAL GRADIENT ACTIVATION MAPPING IN CONVOLUTIONAL NEURAL NETWORKS

被引:3
作者
Intraraprasit, Monthon [1 ]
Chitsobhuk, Orachat [1 ]
机构
[1] King Mongkuts Inst Technol Ladkrabang, Sch Engn, Chalongkrung Rd, Bangkok 10520, Thailand
来源
INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL | 2023年 / 19卷 / 06期
关键词
Filter pruning; Convolutional neural networks; Deep learning; Model com-pression;
D O I
10.24507/ijicic.19.06.1697
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional Neural Network (CNN) is a well-known Deep learning model utilized extensively in the field of computer vision. The structure of convolutional neu-ral networks is quite complicated and necessitates a substantial amount of computational time and storage resources. As a result, it is difficult to adopt a CNN model on a resource -constraint device. Model pruning can help to reduce computation time and storage re-quirements. In this research, we propose a filter pruning technique based on Localized Gradient Activation heatmaP (LGAP) for the purpose of pruning CNNs. Analyzing a filter based on statistical criterion of single neuron can lead to a loss in spatial relations within the filter activation itself, the relationship to target prediction, as well as the re-lationship among filters in that specific layer. To minimize the limitations, we evaluate the significance of a filter through the spatial information of local gradient activation re-lated to the target prediction in terms of the layer-wise loss of the investigated filter. The effect of loss of an investigated filter demonstrates the significance or insignificance of the filter. Our pruning criteria ensure that these significant filters are preserved, while maintaining the model accuracy. The performance of our pruning method was validated using VGG-16 and ResNet-50. With pruning ratio of 50%, VGG-16 tends to decrease 1.66% of its accuracy, 3.6x of FLOP and 3.9x of storage reduction. For ResNet-50, with 50% pruning ratio, the results show that Top-1 and Top-5 of our pruning techniques outperform all the baseline techniques with a reduction of top-1 accuracy by 3.56%, top-5 accuracy by 1.89%, Floating Point Operation by 2.3x, and storage by 2.05x.
引用
收藏
页码:1697 / 1715
页数:19
相关论文
共 38 条
[1]  
[Anonymous], 2016, ARXIV160703250
[2]  
[Anonymous], 1989, P ADV NEUR INF PROC
[3]  
Chollet F., 2015, About us
[4]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[5]   Neuro-fuzzy patch-wise R-CNN for multiple sclerosis segmentation [J].
Essa, Ehab ;
Aldesouky, Doaa ;
Hussein, Sherif E. ;
Rashad, M. Z. .
MEDICAL & BIOLOGICAL ENGINEERING & COMPUTING, 2020, 58 (09) :2161-2175
[6]   A Survey on Efficient Convolutional Neural Networks and Hardware Acceleration [J].
Ghimire, Deepak ;
Kil, Dayoung ;
Kim, Seong-heum .
ELECTRONICS, 2022, 11 (06)
[7]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[8]   Rich feature hierarchies for accurate object detection and semantic segmentation [J].
Girshick, Ross ;
Donahue, Jeff ;
Darrell, Trevor ;
Malik, Jitendra .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :580-587
[9]  
Han S, 2015, ADV NEUR IN, V28
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778