FP-AGL: Filter Pruning With Adaptive Gradient Learning for Accelerating Deep Convolutional Neural Networks

被引:26
作者
Kim, Nam Joon [1 ,2 ]
Kim, Hyun [1 ,2 ]
机构
[1] Seoul Natl Univ Sci & Technol, Dept Elect & Informat Engn, Seoul 01811, South Korea
[2] Seoul Natl Univ Sci & Technol, Res Ctr Elect & Informat Technol, Seoul 01811, South Korea
基金
新加坡国家研究基金会;
关键词
Adaptive gradient learning; convolutional neural networks; filter pruning; light-weight technique; taylor expansion; CNN;
D O I
10.1109/TMM.2022.3189496
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Filter pruning is a technique that reduces computational complexity, inference time, and memory footprint by removing unnecessary filters in convolutional neural networks (CNNs) with an acceptable drop in accuracy, consequently accelerating the network. Unlike traditional filter pruning methods utilizing zeroing-out filters, we propose two techniques to achieve the effect of pruning more filters with less performance degradation, inspired by the existing research on centripetal stochastic gradient descent (C-SGD), wherein the filters are removed only when the ones that need to be pruned have the same value. First, to minimize the negative effect of centripetal vectors that gradually make filters come closer to each other, we redesign the vectors by considering the effect of each vector on the loss-function using the Taylor-based method. Second, we propose an adaptive gradient learning (AGL) technique that updates weights while adaptively changing the gradients. Through AGL, performance degradation can be mitigated because some gradients maintain their original direction, and AGL also minimizes the accuracy loss by perfectly converging the filters, which require pruning, to a single point. Finally, we demonstrate the superiority of the proposed method on various datasets and networks. In particular, on the ILSVRC-2012 dataset, our method removed 52.09% FLOPs with a negligible 0.15% top-1 accuracy drop on ResNet-50. As a result, we achieve the most outstanding performance compared to those reported in previous studies in terms of the trade-off between accuracy and computational complexity.
引用
收藏
页码:5279 / 5290
页数:12
相关论文
共 50 条
[41]   Learning ability of interpolating deep convolutional neural networks [J].
Zhou, Tian-Yi ;
Huo, Xiaoming .
APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2024, 68
[42]   LEARNING CONVOLUTIONAL NEURAL NETWORKS WITH DEEP PART EMBEDDINGS [J].
Gupta, Nitin ;
Mujumdar, Shashank ;
Agarwal, Prerna ;
Jain, Abhinav ;
Mehta, Sameep .
2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, :2037-2041
[43]   Hebbian Learning Meets Deep Convolutional Neural Networks [J].
Amato, Giuseppe ;
Carrara, Fabio ;
Falchi, Fabrizio ;
Gennaro, Claudio ;
Lagani, Gabriele .
IMAGE ANALYSIS AND PROCESSING - ICIAP 2019, PT I, 2019, 11751 :324-334
[44]   SSFP: A Structured Stripe-Filter Pruning Method for Deep Neural Networks [J].
Liu, Jingjing ;
Huang, Lingjin ;
Feng, Manlong ;
Guo, Aiying .
2024 13TH INTERNATIONAL CONFERENCE ON COMMUNICATIONS, CIRCUITS AND SYSTEMS, ICCCAS 2024, 2024, :80-84
[45]   An Updated Survey of Efficient Hardware Architectures for Accelerating Deep Convolutional Neural Networks [J].
Capra, Maurizio ;
Bussolino, Beatrice ;
Marchisio, Alberto ;
Shafique, Muhammad ;
Masera, Guido ;
Martina, Maurizio .
FUTURE INTERNET, 2020, 12 (07)
[46]   Data-Efficient Adaptive Global Pruning for Convolutional Neural Networks in Edge Computing [J].
Gao, Zhipeng ;
Sun, Shan ;
Mo, Zijia ;
Rui, Lanlan ;
Yang, Yang .
ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, :6633-6638
[47]   SAAF: Self-Adaptive Attention Factor-Based Taylor-Pruning on Convolutional Neural Networks [J].
Lu, Yiheng ;
Gong, Maoguo ;
Feng, Kaiyuan ;
Liu, Jialu ;
Guan, Ziyu ;
Li, Hao .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (05) :8540-8553
[48]   Relay Backpropagation for Effective Learning of Deep Convolutional Neural Networks [J].
Shen, Li ;
Lin, Zhouchen ;
Huang, Qingming .
COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 :467-482
[49]   Residual learning of deep convolutional neural networks for image denoising [J].
Shan, Chuanhui ;
Guo, Xirong ;
Ou, Jun .
JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2019, 37 (02) :2809-2818
[50]   Respiratory Sounds Feature Learning with Deep Convolutional Neural Networks [J].
Liu, Yongpeng ;
Lin, Yusong ;
Gao, Shan ;
Zhang, Hongpo ;
Wang, Zongmin ;
Gao, Yang ;
Chen, Guanling .
2017 IEEE 15TH INTL CONF ON DEPENDABLE, AUTONOMIC AND SECURE COMPUTING, 15TH INTL CONF ON PERVASIVE INTELLIGENCE AND COMPUTING, 3RD INTL CONF ON BIG DATA INTELLIGENCE AND COMPUTING AND CYBER SCIENCE AND TECHNOLOGY CONGRESS(DASC/PICOM/DATACOM/CYBERSCI, 2017, :170-177