OPTIMIZING MUSIC SOURCE SEPARATION IN COMPLEX AUDIO ENVIRONMENTS THROUGH PROGRESSIVE SELF-KNOWLEDGE DISTILLATION

被引:0
|
作者
Han, ChangHeon [1 ]
Lee, SuHyun [1 ]
机构
[1] Hanyang Univ, Dept Appl Artificial Intelligence, Seoul, South Korea
来源
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW 2024 | 2024年
关键词
Music Source Separation; Self-Knowledge Distillation; Curriculum Learning; Hearing Loss;
D O I
10.1109/ICASSPW62465.2024.10626965
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This technical report presents our approach for The ICASSP 2024 SP Cadenza Grand Challenge (CADICASSP24), focusing on effective source separation. In the scenario addressed by this challenge, signals captured by hearing aid microphones are complicated by the intertwining of stereo signals, increasing their complexity. In such situations, where simple training losses like L1 loss can lead to significant errors and complicate model training, we introduce an effective fine-tuning method that softens the target using predictions from the previous epoch's model. Our system improved the SDR score by 1.2 dB over the baseline. The source code is available online.
引用
收藏
页码:13 / 14
页数:2
相关论文
共 3 条
  • [1] Self-Knowledge Distillation with Progressive Refinement of Targets
    Kim, Kyungyul
    Ji, ByeongMoon
    Yoon, Doyoung
    Hwang, Sangheum
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 6547 - 6556
  • [2] Self-Knowledge Distillation via Progressive Associative Learning
    Zhao, Haoran
    Bi, Yanxian
    Tian, Shuwen
    Wang, Jian
    Zhang, Peiying
    Deng, Zhaopeng
    Liu, Kai
    ELECTRONICS, 2024, 13 (11)
  • [3] M3-Embedding: Multi-Linguality, Multi-Functionality, Multi-Granularity Text Embeddings Through Self-Knowledge Distillation
    Chen, Jianlv
    Xiao, Shitao
    Zhang, Peitian
    Luo, Kun
    Lian, Defu
    Liu, Zheng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 2318 - 2335