FP-AGL: Filter Pruning With Adaptive Gradient Learning for Accelerating Deep Convolutional Neural Networks

被引:26
作者
Kim, Nam Joon [1 ,2 ]
Kim, Hyun [1 ,2 ]
机构
[1] Seoul Natl Univ Sci & Technol, Dept Elect & Informat Engn, Seoul 01811, South Korea
[2] Seoul Natl Univ Sci & Technol, Res Ctr Elect & Informat Technol, Seoul 01811, South Korea
基金
新加坡国家研究基金会;
关键词
Adaptive gradient learning; convolutional neural networks; filter pruning; light-weight technique; taylor expansion; CNN;
D O I
10.1109/TMM.2022.3189496
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Filter pruning is a technique that reduces computational complexity, inference time, and memory footprint by removing unnecessary filters in convolutional neural networks (CNNs) with an acceptable drop in accuracy, consequently accelerating the network. Unlike traditional filter pruning methods utilizing zeroing-out filters, we propose two techniques to achieve the effect of pruning more filters with less performance degradation, inspired by the existing research on centripetal stochastic gradient descent (C-SGD), wherein the filters are removed only when the ones that need to be pruned have the same value. First, to minimize the negative effect of centripetal vectors that gradually make filters come closer to each other, we redesign the vectors by considering the effect of each vector on the loss-function using the Taylor-based method. Second, we propose an adaptive gradient learning (AGL) technique that updates weights while adaptively changing the gradients. Through AGL, performance degradation can be mitigated because some gradients maintain their original direction, and AGL also minimizes the accuracy loss by perfectly converging the filters, which require pruning, to a single point. Finally, we demonstrate the superiority of the proposed method on various datasets and networks. In particular, on the ILSVRC-2012 dataset, our method removed 52.09% FLOPs with a negligible 0.15% top-1 accuracy drop on ResNet-50. As a result, we achieve the most outstanding performance compared to those reported in previous studies in terms of the trade-off between accuracy and computational complexity.
引用
收藏
页码:5279 / 5290
页数:12
相关论文
共 50 条
[21]   Filter pruning for convolutional neural networks in semantic image segmentation [J].
Lopez-Gonzalez, Clara I. ;
Gasco, Esther ;
Barrientos-Espillco, Fredy ;
Besada-Portas, Eva ;
Pajares, Gonzalo .
NEURAL NETWORKS, 2024, 169 :713-732
[22]   Fpar: filter pruning via attention and rank enhancement for deep convolutional neural networks acceleration [J].
Chen, Yanming ;
Wu, Gang ;
Shuai, Mingrui ;
Lou, Shubin ;
Zhang, Yiwen ;
An, Zhulin .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (07) :2973-2985
[23]   A Gradient Boosting Approach for Training Convolutional and Deep Neural Networks [J].
Emami, Seyedsaman ;
Martinez-Munoz, Gonzalo .
IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2023, 4 :313-321
[24]   Adding Before Pruning: Sparse Filter Fusion for Deep Convolutional Neural Networks via Auxiliary Attention [J].
Tian, Guanzhong ;
Sun, Yiran ;
Liu, Yuang ;
Zeng, Xianfang ;
Wang, Mengmeng ;
Liu, Yong ;
Zhang, Jiangning ;
Chen, Jun .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (03) :3930-3942
[25]   FPC: Filter pruning via the contribution of output feature map for deep convolutional neural networks acceleration [J].
Chen, Yanming ;
Wen, Xiang ;
Zhang, Yiwen ;
He, Qiang .
KNOWLEDGE-BASED SYSTEMS, 2022, 238
[26]   Filter Pruning via Feature Discrimination in Deep Neural Networks [J].
He, Zhiqiang ;
Qian, Yaguan ;
Wang, Yuqi ;
Wang, Bin ;
Guan, Xiaohui ;
Gu, Zhaoquan ;
Ling, Xiang ;
Zeng, Shaoning ;
Wang, Haijiang ;
Zhou, Wujie .
COMPUTER VISION, ECCV 2022, PT XXI, 2022, 13681 :245-261
[27]   Learning sparse convolutional neural networks through filter pruning for efficient fault diagnosis on edge devices [J].
Xu, Gaowei ;
Zhao, Yukai ;
Liu, Min .
NONDESTRUCTIVE TESTING AND EVALUATION, 2025,
[28]   Adaptive Dynamic Filter Pruning Approach Based on Deep Learning [J].
Chu Jinghui ;
Li Meng ;
Lu Wei .
LASER & OPTOELECTRONICS PROGRESS, 2022, 59 (24)
[29]   Self-distillation enhanced adaptive pruning of convolutional neural networks [J].
Diao, Huabin ;
Li, Gongyan ;
Xu, Shaoyun ;
Kong, Chao ;
Wang, Wei ;
Liu, Shuai ;
He, Yuefeng .
PATTERN RECOGNITION, 2025, 157
[30]   V-SKP: Vectorized Kernel-Based Structured Kernel Pruning for Accelerating Deep Convolutional Neural Networks [J].
Koo, Kwanghyun ;
Kim, Hyun .
IEEE ACCESS, 2023, 11 :118547-118557