Adaptive Scaling Filter Pruning Method for Vision Networks With Embedded Devices

被引:0
作者
Ko, Hyunjun [1 ]
Kang, Jin-Ku [1 ]
Kim, Yongwoo [2 ]
机构
[1] Inha Univ, Dept Elect & Comp Engn, Incheon 22212, South Korea
[2] Korea Natl Univ Educ, Dept Technol Educ, Cheongju 28173, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
基金
新加坡国家研究基金会;
关键词
Information filters; Adaptive systems; Adaptive filters; Training; Filtering algorithms; Quantization (signal); Batch normalization; Computer vision; Convolutional neural networks; Deep learning; convolutional neural network; inference time; network compression; pruning;
D O I
10.1109/ACCESS.2024.3454329
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Owing to improvements in computing power, deep learning technology using convolutional neural networks (CNNs) has recently been used in various fields. However, using CNNs on edge devices is challenging because of the large computation required to achieve high performance. To solve this problem, pruning, which reduces redundant parameters and computations, has been widely studied. However, a conventional pruning method requires two learning processes, which are time-consuming and resource-intensive, and it is difficult to reflect the redundancy in the pruned network because it only performs pruning once on the unpruned network. Therefore, in this paper, we utilize a single learning process and propose an adaptive scaling method that dynamically adjusts the size of the network to reflect the changing redundancy in the pruned network. To verify the performance of each method, we compare the performance of the proposed methods by conducting experiments on various datasets and networks. In our experiments using the ImageNet dataset on ResNet-50, pruning FLOPs by 50.1% and 74.0% resulted in a decrease in top-1 accuracy by 0.92% and 3.38%, respectively, and improved inference time by 26.4% and 58.9%, respectively. In addition, pruning FLOPs by 27.37%, 36.84% and 46.41% using the COCO dataset on YOLOv7, reduced mAP(0.5-0.95) by 1.2%, 2.2% and 2.9%, respectively, and improved inference time by 12.9%, 16.9% and19.3%.
引用
收藏
页码:123771 / 123781
页数:11
相关论文
共 40 条
[11]   Data-Driven Sparse Structure Selection for Deep Neural Networks [J].
Huang, Zehao ;
Wang, Naiyan .
COMPUTER VISION - ECCV 2018, PT XVI, 2018, 11220 :317-334
[12]  
Ioffe Sergey, 2015, Proceedings of Machine Learning Research, V37, P448
[13]   Target Capacity Filter Pruning Method for Optimized Inference Time Based on YOLOv5 in Embedded Systems [J].
Jeon, Jihun ;
Kim, Jaemyung ;
Kang, Jin-Ku ;
Moon, Sungtae ;
Kim, Yongwoo .
IEEE ACCESS, 2022, 10 :70840-70849
[14]  
Kang Minsoo, 2020, PR MACH LEARN RES, V119
[15]  
Krizhevsky A, 2009, LEARNING MULTIPLE LA
[16]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[17]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324
[18]  
Li BL, 2020, Img Proc Comp Vis Re, V12347, P639, DOI 10.1007/978-3-030-58536-5_38
[19]  
Li H, 2017, Arxiv, DOI arXiv:1608.08710
[20]   Revisiting Random Channel Pruning for Neural Network Compression [J].
Li, Yawei ;
Adamczewski, Kamil ;
Li, Wen ;
Gu, Shuhang ;
Timofte, Radu ;
Van Gool, Luc .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :191-201