Matching the Ideal Pruning Method with Knowledge Distillation for Optimal Compression

被引:2
|
作者
Malihi, Leila [1 ]
Heidemann, Gunther [1 ]
机构
[1] Osnabruck Univ, Inst Cognit Sci, Dept Comp Vis, D-49074 Osnabruck, Germany
关键词
knowledge distillation; network efficiency; parameter reduction; unstructured pruning; structured pruning;
D O I
10.3390/asi7040056
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, model compression techniques have gained significant attention as a means to reduce the computational and memory requirements of deep neural networks. Knowledge distillation and pruning are two prominent approaches in this domain, each offering unique advantages in achieving model efficiency. This paper investigates the combined effects of knowledge distillation and two pruning strategies, weight pruning and channel pruning, on enhancing compression efficiency and model performance. The study introduces a metric called "Performance Efficiency" to evaluate the impact of these pruning strategies on model compression and performance. Our research is conducted on the popular datasets CIFAR-10 and CIFAR-100. We compared diverse model architectures, including ResNet, DenseNet, EfficientNet, and MobileNet. The results emphasize the efficacy of both weight and channel pruning in achieving model compression. However, a significant distinction emerges, with weight pruning showing superior performance across all four architecture types. We realized that the weight pruning method better adapts to knowledge distillation than channel pruning. Pruned models show a significant reduction in parameters without a significant reduction in accuracy.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Compression of Acoustic Model via Knowledge Distillation and Pruning
    Li, Chenxing
    Zhu, Lei
    Xu, Shuang
    Gao, Peng
    Xu, Bo
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 2785 - 2790
  • [2] Joint structured pruning and dense knowledge distillation for efficient transformer model compression
    Cui, Baiyun
    Li, Yingming
    Zhang, Zhongfei
    NEUROCOMPUTING, 2021, 458 : 56 - 69
  • [3] Effective Compression of Language Models by Combining Pruning and Knowledge Distillation
    Chiu, Chi-Yu
    Hong, Ding-Yong
    Liu, Pangfeng
    Wu, Jan-Jan
    2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024, 2024, : 429 - 438
  • [4] PQK: Model Compression via Pruning, Quantization, and Knowledge Distillation
    Kim, Jangho
    Chang, Simyung
    Kwak, Nojun
    INTERSPEECH 2021, 2021, : 4568 - 4572
  • [5] The Optimization Method of Knowledge Distillation Based on Model Pruning
    Wu, Min
    Ma, Weihua
    Li, Yue
    Zhao, Xiongbo
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 1386 - 1390
  • [6] Efficient and Controllable Model Compression through Sequential Knowledge Distillation and Pruning
    Malihi, Leila
    Heidemann, Gunther
    BIG DATA AND COGNITIVE COMPUTING, 2023, 7 (03)
  • [7] Model compression via pruning and knowledge distillation for person re-identification
    Xie, Haonan
    Jiang, Wei
    Luo, Hao
    Yu, Hongyan
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 12 (02) : 2149 - 2161
  • [8] Model Compression by Iterative Pruning with Knowledge Distillation and Its Application to Speech Enhancement
    Wei, Zeyuan
    Li, Hao
    Zhang, Xueliang
    INTERSPEECH 2022, 2022, : 941 - 945
  • [9] Model compression via pruning and knowledge distillation for person re-identification
    Haonan Xie
    Wei Jiang
    Hao Luo
    Hongyan Yu
    Journal of Ambient Intelligence and Humanized Computing, 2021, 12 : 2149 - 2161
  • [10] A Pruning and Distillation Based Compression Method for Sonar Image Detection Models
    Cheng, Chensheng
    Hou, Xujia
    Wang, Can
    Wen, Xin
    Liu, Weidong
    Zhang, Feihu
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2024, 12 (06)