Matching the Ideal Pruning Method with Knowledge Distillation for Optimal Compression

被引:2
|
作者
Malihi, Leila [1 ]
Heidemann, Gunther [1 ]
机构
[1] Osnabruck Univ, Inst Cognit Sci, Dept Comp Vis, D-49074 Osnabruck, Germany
关键词
knowledge distillation; network efficiency; parameter reduction; unstructured pruning; structured pruning;
D O I
10.3390/asi7040056
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, model compression techniques have gained significant attention as a means to reduce the computational and memory requirements of deep neural networks. Knowledge distillation and pruning are two prominent approaches in this domain, each offering unique advantages in achieving model efficiency. This paper investigates the combined effects of knowledge distillation and two pruning strategies, weight pruning and channel pruning, on enhancing compression efficiency and model performance. The study introduces a metric called "Performance Efficiency" to evaluate the impact of these pruning strategies on model compression and performance. Our research is conducted on the popular datasets CIFAR-10 and CIFAR-100. We compared diverse model architectures, including ResNet, DenseNet, EfficientNet, and MobileNet. The results emphasize the efficacy of both weight and channel pruning in achieving model compression. However, a significant distinction emerges, with weight pruning showing superior performance across all four architecture types. We realized that the weight pruning method better adapts to knowledge distillation than channel pruning. Pruned models show a significant reduction in parameters without a significant reduction in accuracy.
引用
收藏
页数:17
相关论文
共 50 条
  • [41] Boosting Lightweight CNNs Through Network Pruning and Knowledge Distillation for SAR Target Recognition
    Wang, Zhen
    Du, Lan
    Li, Yi
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 : 8386 - 8397
  • [42] Pruning-and-distillation: One-stage joint compression framework for CNNs via clustering
    Niu, Tao
    Teng, Yinglei
    Jin, Lei
    Zou, Panpan
    Liu, Yiding
    IMAGE AND VISION COMPUTING, 2023, 136
  • [43] 3-D Line Matching Network Based on Matching Existence Guidance and Knowledge Distillation
    Tang, Jie
    Liu, Yong
    Yu, Bo
    Liu, Xue
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (20): : 33418 - 33438
  • [44] Optimal Knowledge Distillation through Non-Heuristic Control of Dark Knowledge
    Onchis, Darian
    Istin, Codruta
    Samuila, Ioan
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (03): : 1921 - 1935
  • [45] Skill-Transferring Knowledge Distillation Method
    Yang, Shunzhi
    Xu, Liuchi
    Zhou, Mengchu
    Yang, Xiong
    Yang, Jinfeng
    Huang, Zhenhua
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (11) : 6487 - 6502
  • [46] Defect Detection Method Based on Knowledge Distillation
    Zhou, Qunying
    Wang, Hongyuan
    Tang, Ying
    Wang, Yang
    IEEE ACCESS, 2023, 11 : 35866 - 35873
  • [47] Incremental attribute learning by knowledge distillation method
    Kuang, Zhejun
    Wang, Jingrui
    Sun, Dawen
    Zhao, Jian
    Shi, Lijuan
    Xiong, Xingbo
    JOURNAL OF COMPUTATIONAL DESIGN AND ENGINEERING, 2024, 11 (05) : 259 - 283
  • [48] Counterclockwise block-by-block knowledge distillation for neural network compression
    Lan, Xiaowei
    Zeng, Yalin
    Wei, Xiaoxia
    Zhang, Tian
    Wang, Yiwen
    Huang, Chao
    He, Weikai
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [49] Block-Wisely Supervised Network Pruning with Knowledge Distillation and Markov Chain Monte Carlo
    Liu, Huidong
    Du, Fang
    Song, Lijuan
    Yu, Zhenhua
    APPLIED SCIENCES-BASEL, 2022, 12 (21):
  • [50] Heuristic Compression Method for CNN Model Applying Quantization to a Combination of Structured and Unstructured Pruning Techniques
    Tian, Danhe
    Yamagiwa, Shinichi
    Wada, Koichi
    IEEE ACCESS, 2024, 12 : 66680 - 66689