SKILL: SIMILARITY-AWARE KNOWLEDGE DISTILLATION FOR SPEECH SELF-SUPERVISED LEARNING

被引:0
|
作者
Zampierin, Luca [1 ,2 ]
Hacene, Ghouthi Boukli [1 ,5 ]
Nguyen, Bac [1 ]
Ravanelli, Mirco [3 ,4 ,5 ]
机构
[1] Sony Europe BV, Stuttgart Lab 1, Stuttgart, Germany
[2] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
[3] Concordia Univ, Montreal, PQ, Canada
[4] Univ Montreal, Montreal, PQ, Canada
[5] Mila Quebec AI Inst, Montreal, PQ, Canada
来源
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW 2024 | 2024年
关键词
Model compression; self-supervised learning; knowledge distillation;
D O I
10.1109/ICASSPW62465.2024.10626978
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Self-supervised learning (SSL) has achieved remarkable success across various speech-processing tasks. To enhance its efficiency, previous works often leverage the use of compression techniques. A notable recent attempt is DPHuBERT, which applies joint knowledge distillation (KD) and structured pruning to learn a significantly smaller SSL model. In this paper, we contribute to this research domain by introducing SKILL, a novel method that conducts distillation across groups of layers instead of distilling individual arbitrarily selected layers within the teacher network. The identification of the layers to distill is achieved through a hierarchical clustering procedure applied to layer similarity measures. Extensive experiments demonstrate that our distilled version ofWavLM Base+ not only outperforms DPHuBERT but also achieves state-of-the-art results in the 30M parameters model class across several SUPERB tasks.
引用
收藏
页码:675 / 679
页数:5
相关论文
共 50 条
  • [1] FitHuBERT: Going Thinner and Deeper for Knowledge Distillation of Speech Self-Supervised Learning
    Lee, Yeonghyeon
    Jang, Kangwook
    Goo, Jahyun
    Jung, Youngmoon
    Kim, Hoirin
    INTERSPEECH 2022, 2022, : 3588 - 3592
  • [2] Self-supervised knowledge distillation in counterfactual learning for VQA
    Bi, Yandong
    Jiang, Huajie
    Zhang, Hanfu
    Hu, Yongli
    Yin, Baocai
    PATTERN RECOGNITION LETTERS, 2024, 177 : 33 - 39
  • [3] Self-supervised knowledge distillation for complementary label learning
    Liu, Jiabin
    Li, Biao
    Lei, Minglong
    Shi, Yong
    NEURAL NETWORKS, 2022, 155 : 318 - 327
  • [4] On-Device Constrained Self-Supervised Speech Representation Learning for Keyword Spotting via Knowledge Distillation
    Yang, Gene-Ping
    Gu, Yue
    Tang, Qingming
    Du, Dongsu
    Liu, Yuzong
    INTERSPEECH 2023, 2023, : 1623 - 1627
  • [5] DPHuBERT: Joint Distillation and Pruning of Self-Supervised Speech Models
    Peng, Yifan
    Sudo, Yui
    Muhammad, Shakeel
    Watanabe, Shinji
    INTERSPEECH 2023, 2023, : 62 - 66
  • [6] Image quality assessment based on self-supervised learning and knowledge distillation
    Sang, Qingbing
    Shu, Ziru
    Liu, Lixiong
    Hu, Cong
    Wu, Qin
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 90
  • [7] Knowledge-Aware Graph Self-Supervised Learning for Recommendation
    Li, Shanshan
    Jia, Yutong
    Wu, You
    Wei, Ning
    Zhang, Liyan
    Guo, Jingfeng
    ELECTRONICS, 2023, 12 (23)
  • [8] Hierarchical Self-Supervised Learning for Knowledge-Aware Recommendation
    Zhou, Cong
    Zhou, Sihang
    Huang, Jian
    Wang, Dong
    APPLIED SCIENCES-BASEL, 2024, 14 (20):
  • [9] SIMILARITY ANALYSIS OF SELF-SUPERVISED SPEECH REPRESENTATIONS
    Chung, Yu-An
    Belinkov, Yonatan
    Glass, James
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3040 - 3044
  • [10] Self-Supervised Contrastive Learning for Camera-to-Radar Knowledge Distillation
    Wang, Wenpeng
    Campbell, Bradford
    Munir, Sirajum
    2024 20TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING IN SMART SYSTEMS AND THE INTERNET OF THINGS, DCOSS-IOT 2024, 2024, : 154 - 161