FP-AGL: Filter Pruning With Adaptive Gradient Learning for Accelerating Deep Convolutional Neural Networks

被引:17
作者
Kim, Nam Joon [1 ,2 ]
Kim, Hyun [1 ,2 ]
机构
[1] Seoul Natl Univ Sci & Technol, Dept Elect & Informat Engn, Seoul 01811, South Korea
[2] Seoul Natl Univ Sci & Technol, Res Ctr Elect & Informat Technol, Seoul 01811, South Korea
基金
新加坡国家研究基金会;
关键词
Adaptive gradient learning; convolutional neural networks; filter pruning; light-weight technique; taylor expansion; CNN;
D O I
10.1109/TMM.2022.3189496
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Filter pruning is a technique that reduces computational complexity, inference time, and memory footprint by removing unnecessary filters in convolutional neural networks (CNNs) with an acceptable drop in accuracy, consequently accelerating the network. Unlike traditional filter pruning methods utilizing zeroing-out filters, we propose two techniques to achieve the effect of pruning more filters with less performance degradation, inspired by the existing research on centripetal stochastic gradient descent (C-SGD), wherein the filters are removed only when the ones that need to be pruned have the same value. First, to minimize the negative effect of centripetal vectors that gradually make filters come closer to each other, we redesign the vectors by considering the effect of each vector on the loss-function using the Taylor-based method. Second, we propose an adaptive gradient learning (AGL) technique that updates weights while adaptively changing the gradients. Through AGL, performance degradation can be mitigated because some gradients maintain their original direction, and AGL also minimizes the accuracy loss by perfectly converging the filters, which require pruning, to a single point. Finally, we demonstrate the superiority of the proposed method on various datasets and networks. In particular, on the ILSVRC-2012 dataset, our method removed 52.09% FLOPs with a negligible 0.15% top-1 accuracy drop on ResNet-50. As a result, we achieve the most outstanding performance compared to those reported in previous studies in terms of the trade-off between accuracy and computational complexity.
引用
收藏
页码:5279 / 5290
页数:12
相关论文
共 50 条
  • [21] A Gradient Boosting Approach for Training Convolutional and Deep Neural Networks
    Emami, Seyedsaman
    Martinez-Munoz, Gonzalo
    IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2023, 4 : 313 - 321
  • [22] Adding Before Pruning: Sparse Filter Fusion for Deep Convolutional Neural Networks via Auxiliary Attention
    Tian, Guanzhong
    Sun, Yiran
    Liu, Yuang
    Zeng, Xianfang
    Wang, Mengmeng
    Liu, Yong
    Zhang, Jiangning
    Chen, Jun
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021,
  • [23] FPC: Filter pruning via the contribution of output feature map for deep convolutional neural networks acceleration
    Chen, Yanming
    Wen, Xiang
    Zhang, Yiwen
    He, Qiang
    KNOWLEDGE-BASED SYSTEMS, 2022, 238
  • [24] Filter Pruning via Feature Discrimination in Deep Neural Networks
    He, Zhiqiang
    Qian, Yaguan
    Wang, Yuqi
    Wang, Bin
    Guan, Xiaohui
    Gu, Zhaoquan
    Ling, Xiang
    Zeng, Shaoning
    Wang, Haijiang
    Zhou, Wujie
    COMPUTER VISION, ECCV 2022, PT XXI, 2022, 13681 : 245 - 261
  • [25] Adaptive Dynamic Filter Pruning Approach Based on Deep Learning
    Chu Jinghui
    Li Meng
    Lu Wei
    LASER & OPTOELECTRONICS PROGRESS, 2022, 59 (24)
  • [26] Learning sparse convolutional neural networks through filter pruning for efficient fault diagnosis on edge devices
    Xu, Gaowei
    Zhao, Yukai
    Liu, Min
    NONDESTRUCTIVE TESTING AND EVALUATION, 2025,
  • [27] Self-distillation enhanced adaptive pruning of convolutional neural networks
    Diao, Huabin
    Li, Gongyan
    Xu, Shaoyun
    Kong, Chao
    Wang, Wei
    Liu, Shuai
    He, Yuefeng
    PATTERN RECOGNITION, 2025, 157
  • [28] V-SKP: Vectorized Kernel-Based Structured Kernel Pruning for Accelerating Deep Convolutional Neural Networks
    Koo, Kwanghyun
    Kim, Hyun
    IEEE ACCESS, 2023, 11 : 118547 - 118557
  • [29] Neuroplasticity-Based Pruning Method for Deep Convolutional Neural Networks
    Camacho, Jose David
    Villasenor, Carlos
    Lopez-Franco, Carlos
    Arana-Daniel, Nancy
    APPLIED SCIENCES-BASEL, 2022, 12 (10):
  • [30] Joint filter and channel pruning of convolutional neural networks as a bi-level optimization problem
    Hassen Louati
    Ali Louati
    Slim Bechikh
    Elham Kariri
    Memetic Computing, 2024, 16 : 71 - 90