PQK: Model Compression via Pruning, Quantization, and Knowledge Distillation

被引:9
|
作者
Kim, Jangho [1 ,2 ]
Chang, Simyung [1 ]
Kwak, Nojun [2 ]
机构
[1] Qualcomm Korea YH, Qualcomm AI Res, Seoul, South Korea
[2] Seoul Natl Univ, Seoul, South Korea
来源
INTERSPEECH 2021 | 2021年
基金
新加坡国家研究基金会;
关键词
keyword spotting; model pruning; model quantization; knowledge distillation;
D O I
10.21437/Interspeech.2021-248
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
As edge devices become prevalent, deploying Deep Neural Networks (DNN) on edge devices has become a critical issue. However, DNN requires a high computational resource which is rarely available for edge devices. To handle this, we propose a novel model compression method for the devices with limited computational resources, called PQK consisting of pruning, quantization, and knowledge distillation (KD) processes. Unlike traditional pruning and KD, PQK makes use of unimportant weights pruned in the pruning process to make a teacher network for training a better student network without pre-training the teacher model. PQK has two phases. Phase 1 exploits iterative pruning and quantization-aware training to make a lightweight and power-efficient model. In phase 2, we make a teacher network by adding unimportant weights unused in phase 1 to a pruned network. By using this teacher network, we train the pruned network as a student network. In doing so, we do not need a pre-trained teacher network for the KD framework because the teacher and the student networks coexist within the same network (See Fig. 1). We apply our method to the recognition model and verify the effectiveness of PQK on keyword spotting (KWS) and image recognition.
引用
收藏
页码:4568 / 4572
页数:5
相关论文
共 50 条
  • [41] DISCOVER THE EFFECTIVE STRATEGY FOR FACE RECOGNITION MODEL COMPRESSION BY IMPROVED KNOWLEDGE DISTILLATION
    Wang, Mengjiao
    Liu, Rujie
    Abe, Narishige
    Uchida, Hidetsugu
    Matsunami, Tomoaki
    Yamada, Shigefumi
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 2416 - 2420
  • [42] AUGMENTING KNOWLEDGE DISTILLATION WITH PEER-TO-PEER MUTUAL LEARNING FOR MODEL COMPRESSION
    Niyaz, Usma
    Bathula, Deepti R.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (IEEE ISBI 2022), 2022,
  • [43] TGNet: A Lightweight Infrared Thermal Image Gesture Recognition Network Based on Knowledge Distillation and Model Pruning
    Chen, L.
    Sun, Q.
    Xu, Z.
    Liao, Y.
    2024 CROSS STRAIT RADIO SCIENCE AND WIRELESS TECHNOLOGY CONFERENCE, CSRSWTC 2024, 2024, : 96 - 98
  • [44] Incremental event detection via an improved knowledge distillation based model
    Lin, Yi
    Xu, Changhua
    Yu, Hang
    Tian, Pinzhuo
    Luo, Xiangfeng
    NEUROCOMPUTING, 2023, 551
  • [45] Robustness-Reinforced Knowledge Distillation With Correlation Distance and Network Pruning
    Kim, Seonghak
    Ham, Gyeongdo
    Cho, Yucheol
    Kim, Daeshik
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 9163 - 9175
  • [46] Attention-Fused CNN Model Compression with Knowledge Distillation for Brain Tumor Segmentation
    Xu, Pengcheng
    Kim, Kyungsang
    Liu, Huafeng
    Li, Quanzheng
    MEDICAL IMAGE UNDERSTANDING AND ANALYSIS, MIUA 2022, 2022, 13413 : 328 - 338
  • [47] IoT Device Friendly and Communication-Efficient Federated Learning via Joint Model Pruning and Quantization
    Prakash, Pavana
    Ding, Jiahao
    Chen, Rui
    Qin, Xiaoqi
    Shu, Minglei
    Cui, Qimei
    Guo, Yuanxiong
    Pan, Miao
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (15): : 13638 - 13650
  • [48] Data-Free Low-Bit Quantization via Dynamic Multi-teacher Knowledge Distillation
    Huang, Chong
    Lin, Shaohui
    Zhang, Yan
    Li, Ke
    Zhang, Baochang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VIII, 2024, 14432 : 28 - 41
  • [49] Collaborative knowledge distillation via filter knowledge transfer
    Gou, Jianping
    Hu, Yue
    Sun, Liyuan
    Wang, Zhi
    Ma, Hongxing
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [50] Knowledge Distillation via Information Matching
    Zhu, Honglin
    Jiang, Ning
    Tang, Jialiang
    Huang, Xinlei
    NEURAL INFORMATION PROCESSING, ICONIP 2023, PT IV, 2024, 14450 : 405 - 417