Multi-objective evolutionary architectural pruning of deep convolutional neural networks with weights inheritance

被引:3
作者
Chung, K. T. [1 ]
Lee, C. K. M. [1 ,2 ]
Tsang, Y. P. [1 ]
Wu, C. H. [3 ]
Asadipour, Ali [4 ]
机构
[1] Hong Kong Polytech Univ, Dept Ind & Syst Engn, Hong Kong, Peoples R China
[2] Lab Artificial Intelligence Design, Hong Kong Sci Pk, Hong Kong, Peoples R China
[3] Hang Seng Univ Hong Kong, Dept Supply Chain & Informat Management, Hong Kong, Peoples R China
[4] Royal Coll Art, Comp Sci Res Ctr, London, England
关键词
Network pruning; Multi-objective evolutionary algorithm; Deep convolutional neural networks; Deep learning; OPTIMIZATION;
D O I
10.1016/j.ins.2024.121265
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Despite the ongoing success of artificial intelligence applications, the deployment of deep learning models on end devices remains challenging due to the limited onboard computational resources. A way to tackle this challenge is model compression through network pruning, which removes unnecessary parameters to reduce model size without significantly affecting performance. However, existing iterative methods require designated pruning rates and obtain a single pruned model, which lacks the flexibility to adapt to devices with heterogeneous computational capabilities. This paper considers network pruning in Deep Convolutional Neural Networks (DCNNs) and proposes a novel algorithm for structured filter pruning in DCNNs using a multi- objective evolutionary approach with a novel weights inheritance scheme and representation scheme to reduce the time cost of the optimization process. The proposed method provides solutions with multiple levels of tradeoff between performance and efficiency for various hardware specifications on edge devices. Experimental results demonstrate the effectiveness of the proposed framework in optimizing popular DCNN models in terms of model complexity and accuracy. Notably, the framework successfully made significant reductions in floating-point operations ranging from 40% to 90% of VGG-16/19 and ResNet-56/110 with negligible loss in accuracy on the CIFAR-10/100 dataset.
引用
收藏
页数:18
相关论文
共 50 条
[1]   Compression and acceleration of convolution neural network: a Genetic Algorithm based approach [J].
Agarwal, Mohit ;
Gupta, Suneet K. ;
Biswas, Mainak ;
Garg, Deepak .
JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2022, 14 (10) :13387-13397
[2]  
Allen-Zhu Z, 2019, PR MACH LEARN RES, V97
[3]   Pruning CNN filters via quantifying the importance of deep visual representations [J].
Alqahtani, Ali ;
Xie, Xianghua ;
Jones, Mark W. ;
Essa, Ehab .
COMPUTER VISION AND IMAGE UNDERSTANDING, 2021, 208
[4]   Survey of Deep Reinforcement Learning for Motion Planning of Autonomous Vehicles [J].
Aradi, Szilard .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (02) :740-759
[5]  
Benbaki Riade, P MACHINE LEARNING R
[6]   Deep Learning With Edge Computing: A Review [J].
Chen, Jiasi ;
Ran, Xukan .
PROCEEDINGS OF THE IEEE, 2019, 107 (08) :1655-1674
[7]   CCPrune: Collaborative channel pruning for learning compact convolutional networks [J].
Chen, Yanming ;
Wen, Xiang ;
Zhang, Yiwen ;
Shi, Weisong .
NEUROCOMPUTING, 2021, 451 :35-45
[8]  
Cheng HR, 2024, Arxiv, DOI [arXiv:2308.06767, 10.48550/arXiv.2308.06767]
[9]   Heuristic-based automatic pruning of deep neural networks [J].
Choudhary, Tejalal ;
Mishra, Vipul ;
Goswami, Anurag ;
Sarangapani, Jagannathan .
NEURAL COMPUTING & APPLICATIONS, 2022, 34 (06) :4889-4903
[10]   An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point-Based Nondominated Sorting Approach, Part I: Solving Problems With Box Constraints [J].
Deb, Kalyanmoy ;
Jain, Himanshu .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2014, 18 (04) :577-601