Matching the Ideal Pruning Method with Knowledge Distillation for Optimal Compression

被引:2
|
作者
Malihi, Leila [1 ]
Heidemann, Gunther [1 ]
机构
[1] Osnabruck Univ, Inst Cognit Sci, Dept Comp Vis, D-49074 Osnabruck, Germany
关键词
knowledge distillation; network efficiency; parameter reduction; unstructured pruning; structured pruning;
D O I
10.3390/asi7040056
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, model compression techniques have gained significant attention as a means to reduce the computational and memory requirements of deep neural networks. Knowledge distillation and pruning are two prominent approaches in this domain, each offering unique advantages in achieving model efficiency. This paper investigates the combined effects of knowledge distillation and two pruning strategies, weight pruning and channel pruning, on enhancing compression efficiency and model performance. The study introduces a metric called "Performance Efficiency" to evaluate the impact of these pruning strategies on model compression and performance. Our research is conducted on the popular datasets CIFAR-10 and CIFAR-100. We compared diverse model architectures, including ResNet, DenseNet, EfficientNet, and MobileNet. The results emphasize the efficacy of both weight and channel pruning in achieving model compression. However, a significant distinction emerges, with weight pruning showing superior performance across all four architecture types. We realized that the weight pruning method better adapts to knowledge distillation than channel pruning. Pruned models show a significant reduction in parameters without a significant reduction in accuracy.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] Optimizing BERT for Bengali Emotion Classification: Evaluating Knowledge Distillation, Pruning, and Quantization
    Rahman, Md Hasibur
    Uddin, Mohammed Arif
    Ria, Zinnat Fowzia
    Rahman, Rashedur M.
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2025, 142 (02): : 1637 - 1666
  • [32] Knowledge Distillation via Hierarchical Matching for Small Object Detection
    Ma, Yong-Chi
    Ma, Xiao
    Hao, Tian-Ran
    Cui, Li-Sha
    Jin, Shao-Hui
    Lyu, Pei
    Journal of Computer Science and Technology, 2024, 39 (04) : 798 - 810
  • [33] AFMPM: adaptive feature map pruning method based on feature distillation
    Yufeng Guo
    Weiwei Zhang
    Junhuang Wang
    Ming Ji
    Chenghui Zhen
    Zhengzheng Guo
    International Journal of Machine Learning and Cybernetics, 2024, 15 : 573 - 588
  • [34] Model Compression Based on Knowledge Distillation and Its Application in HRRP
    Chen, Xiaojiao
    An, Zhenyu
    Huang, Liansheng
    He, Shiying
    Wang, Zhen
    PROCEEDINGS OF 2020 IEEE 4TH INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2020), 2020, : 1268 - 1272
  • [35] AFMPM: adaptive feature map pruning method based on feature distillation
    Guo, Yufeng
    Zhang, Weiwei
    Wang, Junhuang
    Ji, Ming
    Zhen, Chenghui
    Guo, Zhengzheng
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (02) : 573 - 588
  • [36] DDK: Dynamic structure pruning based on differentiable search and recursive knowledge distillation for BERT
    Zhang, Zhou
    Lu, Yang
    Wang, Tengfei
    Wei, Xing
    Wei, Zhen
    NEURAL NETWORKS, 2024, 173
  • [37] A neural network compression method based on knowledge-distillation and parameter quantization for the bearing fault diagnosis
    Ji, Mengyu
    Peng, Gaoliang
    Li, Sijue
    Cheng, Feng
    Chen, Zhao
    Li, Zhixiong
    Du, Haiping
    APPLIED SOFT COMPUTING, 2022, 127
  • [38] Uncertainty-Driven Knowledge Distillation for Language Model Compression
    Huang, Tianyu
    Dong, Weisheng
    Wu, Fangfang
    Li, Xin
    Shi, Guangming
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 2850 - 2858
  • [39] Parallel Blockwise Knowledge Distillation for Deep Neural Network Compression
    Blakeney, Cody
    Li, Xiaomin
    Yan, Yan
    Zong, Ziliang
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2021, 32 (07) : 1765 - 1776
  • [40] Feature Pruning and Recovery Learning with Knowledge Distillation for Occluded Person Re-Identification
    Hou, Mengyu
    Gan, Wenjun
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VIII, 2025, 15038 : 339 - 353