SKILL: SIMILARITY-AWARE KNOWLEDGE DISTILLATION FOR SPEECH SELF-SUPERVISED LEARNING

被引:0
|
作者
Zampierin, Luca [1 ,2 ]
Hacene, Ghouthi Boukli [1 ,5 ]
Nguyen, Bac [1 ]
Ravanelli, Mirco [3 ,4 ,5 ]
机构
[1] Sony Europe BV, Stuttgart Lab 1, Stuttgart, Germany
[2] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
[3] Concordia Univ, Montreal, PQ, Canada
[4] Univ Montreal, Montreal, PQ, Canada
[5] Mila Quebec AI Inst, Montreal, PQ, Canada
来源
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW 2024 | 2024年
关键词
Model compression; self-supervised learning; knowledge distillation;
D O I
10.1109/ICASSPW62465.2024.10626978
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Self-supervised learning (SSL) has achieved remarkable success across various speech-processing tasks. To enhance its efficiency, previous works often leverage the use of compression techniques. A notable recent attempt is DPHuBERT, which applies joint knowledge distillation (KD) and structured pruning to learn a significantly smaller SSL model. In this paper, we contribute to this research domain by introducing SKILL, a novel method that conducts distillation across groups of layers instead of distilling individual arbitrarily selected layers within the teacher network. The identification of the layers to distill is achieved through a hierarchical clustering procedure applied to layer similarity measures. Extensive experiments demonstrate that our distilled version ofWavLM Base+ not only outperforms DPHuBERT but also achieves state-of-the-art results in the 30M parameters model class across several SUPERB tasks.
引用
收藏
页码:675 / 679
页数:5
相关论文
共 50 条
  • [31] NOISE ROBUST DISTILLATION OF SELF-SUPERVISED SPEECH MODELS VIA CORRELATION METRICS
    Ritter-Gutierrez, Fabian
    Huang, Kuan-Po
    Ng, Dianwen
    Wong, Jeremy H. M.
    Lee, Hung-yi
    Chng, Eng Siong
    Chen, Nancy F.
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW 2024, 2024, : 495 - 499
  • [32] OAGknow: Self-Supervised Learning for Linking Knowledge Graphs
    Liu, Xiao
    Mian, Li
    Dong, Yuxiao
    Zhang, Fanjin
    Zhang, Jing
    Tang, Jie
    Zhang, Peng
    Gong, Jibing
    Wang, Kuansan
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (02) : 1895 - 1908
  • [33] Multi-student Collaborative Self-supervised Distillation
    Yang, Yinan
    Chen, Li
    Wu, Shaohui
    Sun, Zhuang
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT II, 2023, 14087 : 199 - 210
  • [34] Functional Knowledge Transfer with Self-supervised Representation Learning
    Chhipa, Prakash Chandra
    Chopra, Muskaan
    Mengi, Gopal
    Gupta, Varun
    Upadhyay, Richa
    Chippa, Meenakshi Subhash
    De, Kanjar
    Saini, Rajkumar
    Uchida, Seiichi
    Liwicki, Marcus
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3339 - 3343
  • [35] Augmentation-aware self-supervised learning with conditioned projector
    Przewiezlikowski, Marcin
    Pyla, Mateusz
    Zielinski, Bartosz
    Twardowski, Bartlomiej
    Tabor, Jacek
    Smieja, Marek
    KNOWLEDGE-BASED SYSTEMS, 2024, 305
  • [36] Efficient Personalized Speech Enhancement Through Self-Supervised Learning
    Sivaraman, Aswin
    Kim, Minje
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2022, 16 (06) : 1342 - 1356
  • [37] Automatic self-supervised learning of associations between speech and text
    Knuuttila, Juho
    Rasanen, Okko
    Laine, Unto K.
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 465 - 469
  • [38] Linear-Complexity Self-Supervised Learning for Speech Processing
    Zhang, Shucong
    Parcollet, Titouan
    van Dalen, Rogier
    Bhattacharya, Sourav
    INTERSPEECH 2024, 2024, : 3480 - 3484
  • [39] AeroRec: An Efficient On-Device Recommendation Framework using Federated Self-Supervised Knowledge Distillation
    Xia, Tengxi
    Ren, Ju
    Rao, Wei
    Zu, Qin
    Wang, Wenjie
    Chen, Shuai
    Zhang, Yaoxue
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS, 2024, : 121 - 130
  • [40] Traffic Prediction with Self-Supervised Learning: A Heterogeneity-Aware Model for Urban Traffic Flow Prediction Based on Self-Supervised Learning
    Gao, Min
    Wei, Yingmei
    Xie, Yuxiang
    Zhang, Yitong
    MATHEMATICS, 2024, 12 (09)