FP-AGL: Filter Pruning With Adaptive Gradient Learning for Accelerating Deep Convolutional Neural Networks

被引:17
作者
Kim, Nam Joon [1 ,2 ]
Kim, Hyun [1 ,2 ]
机构
[1] Seoul Natl Univ Sci & Technol, Dept Elect & Informat Engn, Seoul 01811, South Korea
[2] Seoul Natl Univ Sci & Technol, Res Ctr Elect & Informat Technol, Seoul 01811, South Korea
基金
新加坡国家研究基金会;
关键词
Adaptive gradient learning; convolutional neural networks; filter pruning; light-weight technique; taylor expansion; CNN;
D O I
10.1109/TMM.2022.3189496
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Filter pruning is a technique that reduces computational complexity, inference time, and memory footprint by removing unnecessary filters in convolutional neural networks (CNNs) with an acceptable drop in accuracy, consequently accelerating the network. Unlike traditional filter pruning methods utilizing zeroing-out filters, we propose two techniques to achieve the effect of pruning more filters with less performance degradation, inspired by the existing research on centripetal stochastic gradient descent (C-SGD), wherein the filters are removed only when the ones that need to be pruned have the same value. First, to minimize the negative effect of centripetal vectors that gradually make filters come closer to each other, we redesign the vectors by considering the effect of each vector on the loss-function using the Taylor-based method. Second, we propose an adaptive gradient learning (AGL) technique that updates weights while adaptively changing the gradients. Through AGL, performance degradation can be mitigated because some gradients maintain their original direction, and AGL also minimizes the accuracy loss by perfectly converging the filters, which require pruning, to a single point. Finally, we demonstrate the superiority of the proposed method on various datasets and networks. In particular, on the ILSVRC-2012 dataset, our method removed 52.09% FLOPs with a negligible 0.15% top-1 accuracy drop on ResNet-50. As a result, we achieve the most outstanding performance compared to those reported in previous studies in terms of the trade-off between accuracy and computational complexity.
引用
收藏
页码:5279 / 5290
页数:12
相关论文
共 50 条
  • [31] Joint filter and channel pruning of convolutional neural networks as a bi-level optimization problem
    Louati, Hassen
    Louati, Ali
    Bechikh, Slim
    Kariri, Elham
    [J]. MEMETIC COMPUTING, 2024, 16 (01) : 71 - 90
  • [32] Hardware-Aware Evolutionary Explainable Filter Pruning for Convolutional Neural Networks
    Christian Heidorn
    Muhammad Sabih
    Nicolai Meyerhöfer
    Christian Schinabeck
    Jürgen Teich
    Frank Hannig
    [J]. International Journal of Parallel Programming, 2024, 52 : 40 - 58
  • [33] A Dual Rank-Constrained Filter Pruning Approach for Convolutional Neural Networks
    Fan, Fugui
    Su, Yuting
    Jing, Peiguang
    Lu, Wei
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2021, 28 (28) : 1734 - 1738
  • [34] Hardware-Aware Evolutionary Explainable Filter Pruning for Convolutional Neural Networks
    Heidorn, Christian
    Sabih, Muhammad
    Meyerhoefer, Nicolai
    Schinabeck, Christian
    Teich, Juergen
    Hannig, Frank
    [J]. INTERNATIONAL JOURNAL OF PARALLEL PROGRAMMING, 2024, 52 (1-2) : 40 - 58
  • [35] Zero-Keep Filter Pruning for Energy/Power Efficient Deep Neural Networks
    Woo, Yunhee
    Kim, Dongyoung
    Jeong, Jaemin
    Ko, Young-Woong
    Lee, Jeong-Gun
    [J]. ELECTRONICS, 2021, 10 (11)
  • [36] SFP: Similarity-based filter pruning for deep neural networks
    Li, Guoqing
    Li, Rengang
    Li, Tuo
    Shen, Chaoyao
    Zou, Xiaofeng
    Wang, Jiuyang
    Wang, Changhong
    Li, Nanjun
    [J]. INFORMATION SCIENCES, 2025, 689
  • [37] Deep imitation learning for autonomous vehicles based on convolutional neural networks
    Kebria, Parham M.
    Khosravi, Abbas
    Salaken, Syed Moshfeq
    Nahavandi, Saeid
    [J]. IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2020, 7 (01) : 82 - 95
  • [38] Learning ability of interpolating deep convolutional neural networks
    Zhou, Tian-Yi
    Huo, Xiaoming
    [J]. APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2024, 68
  • [39] LEARNING CONVOLUTIONAL NEURAL NETWORKS WITH DEEP PART EMBEDDINGS
    Gupta, Nitin
    Mujumdar, Shashank
    Agarwal, Prerna
    Jain, Abhinav
    Mehta, Sameep
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2037 - 2041
  • [40] Deep learning for steganalysis via convolutional neural networks
    Qian, Yinlong
    Dong, Jing
    Wang, Wei
    Tan, Tieniu
    [J]. MEDIA WATERMARKING, SECURITY, AND FORENSICS 2015, 2015, 9409