Efficient Joint Optimization of Layer-Adaptive Weight Pruning in Deep Neural Networks

被引:7
|
作者
Xu, Kaixin [1 ,2 ]
Wang, Zhe [1 ,2 ]
Geng, Xue [1 ]
Wu, Min [1 ]
Li, Xiaoli [1 ,2 ]
Lin, Weisi [2 ]
机构
[1] ASTAR, Inst Infocomm Res I2R, 1 Fusionopolis Way, Singapore 138632, Singapore
[2] Nanyang Technol Univ, Singapore, Singapore
来源
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023) | 2023年
关键词
D O I
10.1109/ICCV51070.2023.01600
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a novel layer-adaptive weight-pruning approach for Deep Neural Networks (DNNs) that addresses the challenge of optimizing the output distortion minimization while adhering to a target pruning ratio constraint. Our approach takes into account the collective influence of all layers to design a layer-adaptive pruning scheme. We discover and utilize a very important additivity property of output distortion caused by pruning weights on multiple layers. This property enables us to formulate the pruning as a combinatorial optimization problem and efficiently solve it through dynamic programming. By decomposing the problem into sub-problems, we achieve linear time complexity, making our optimization algorithm fast and feasible to run on CPUs. Our extensive experiments demonstrate the superiority of our approach over existing methods on the ImageNet and CIFAR-10 datasets. On CIFAR-10, our method achieves remarkable improvements, outperforming others by up to 1.0% for ResNet-32, 0.5% for VGG-16, and 0.7% for DenseNet-121 in terms of top-1 accuracy. On ImageNet, we achieve up to 4.7% and 4.6% higher top-1 accuracy compared to other methods for VGG16 and ResNet-50, respectively. These results highlight the effectiveness and practicality of our approach for enhancing DNN performance through layer-adaptive weight pruning. Code will be available on https://github.com/Akimoto-Cris/RD_VIT_PRUNE.
引用
收藏
页码:17401 / 17411
页数:11
相关论文
共 50 条
  • [41] Activation Pruning of Deep Convolutional Neural Networks
    Ardakani, Arash
    Condo, Carlo
    Gross, Warren J.
    2017 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2017), 2017, : 1325 - 1329
  • [42] Fast Convex Pruning of Deep Neural Networks
    Aghasi, Alireza
    Abdi, Afshin
    Romberg, Justin
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2020, 2 (01): : 158 - 188
  • [43] LAYER-WISE DEEP NEURAL NETWORK PRUNING VIA ITERATIVELY REWEIGHTED OPTIMIZATION
    Jiang, Tao
    Yang, Xiangyu
    Shi, Yuanming
    Wang, Hao
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 5606 - 5610
  • [44] Optimizing the Deep Neural Networks by Layer-Wise Refined Pruning and the Acceleration on FPGA
    Li, Hengyi
    Yue, Xuebin
    Wang, Zhichen
    Chai, Zhilei
    Wang, Wenwen
    Tomiyama, Hiroyuki
    Meng, Lin
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [45] Once For All Skip: Efficient Adaptive Deep Neural Networks
    Yang, Yu
    Liu, Di
    Fang, Hui
    Huang, Yi-Xiong
    Sun, Ying
    Zhang, Zhi-Yuan
    PROCEEDINGS OF THE 2022 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2022), 2022, : 568 - 571
  • [46] Data-Efficient Adaptive Global Pruning for Convolutional Neural Networks in Edge Computing
    Gao, Zhipeng
    Sun, Shan
    Mo, Zijia
    Rui, Lanlan
    Yang, Yang
    ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 6633 - 6638
  • [47] Efficient out-of-distribution detection via layer-adaptive scoring and early stopping
    Wang, Haoliang
    Zhao, Chen
    Chen, Feng
    FRONTIERS IN BIG DATA, 2024, 7
  • [48] Sparse optimization guided pruning for neural networks
    Shi, Yong
    Tang, Anda
    Niu, Lingfeng
    Zhou, Ruizhi
    NEUROCOMPUTING, 2024, 574
  • [49] Automatic Pruning Rate Derivation for Structured Pruning of Deep Neural Networks
    Sakai, Yasufumi
    Iwakawa, Akinori
    Tabaru, Tsuguchika
    Inoue, Atsuki
    Kawaguchi, Hiroshi
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2561 - 2567
  • [50] Zero-Keep Filter Pruning for Energy/Power Efficient Deep Neural Networks
    Woo, Yunhee
    Kim, Dongyoung
    Jeong, Jaemin
    Ko, Young-Woong
    Lee, Jeong-Gun
    ELECTRONICS, 2021, 10 (11)