共 50 条
- [1] AdaDS: Adaptive data selection for accelerating pre-trained language model knowledge distillation AI OPEN, 2023, 4 : 56 - 63
- [3] ARC: A Layer Replacement Compression Method Based on Fine-Grained Self-Attention Distillation for Compressing Pre-Trained Language Models IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2025, 9 (01): : 848 - 860
- [4] GenDistiller: Distilling Pre-trained Language Models based on an Autoregressive Generative Model INTERSPEECH 2024, 2024, : 3325 - 3329
- [5] Knowledge Transfer from Pre-trained Language Models to Cif-based Speech Recognizers via Hierarchical Distillation INTERSPEECH 2023, 2023, : 1364 - 1368
- [7] KNOWLEDGE DISTILLATION FOR NEURAL TRANSDUCERS FROM LARGE SELF-SUPERVISED PRE-TRAINED MODELS 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8527 - 8531
- [9] A Light Bug Triage Framework for Applying Large Pre-trained Language Model PROCEEDINGS OF THE 37TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE 2022, 2022,
- [10] Fast and accurate image retrieval using knowledge distillation from multiple deep pre-trained networks Multimedia Tools and Applications, 2023, 82 : 33937 - 33959