DISCRIMINATIVE DEEP RECURRENT NEURAL NETWORKS FOR MONAURAL SPEECH SEPARATION

被引:0
|
作者
Wang, Guan-Xiang [1 ]
Hsu, Chung-Chien [1 ]
Chien, Jen-Tzung [1 ]
机构
[1] Natl Chiao Tung Univ, Dept Elect & Comp Engn, Hsinchu 30010, Taiwan
关键词
deep learning; discriminative learning; neural network; monaural speech separation; FACTORIZATION;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Deep neural network is now a new trend towards solving different problems in speech processing. In this paper, we propose a discriminative deep recurrent neural network (DRNN) model for monaural speech separation. Our idea is to construct DRNN as a regression model to discover the deep structure and regularity for signal reconstruction from a mixture of two source spectra. To reinforce the discrimination capability between two separated spectra, we estimate DRNN separation parameters by minimizing an integrated objective function which consists of two measurements. One is the within source reconstruction errors due to the individual source spectra while the other conveys the discrimination information which preserves the mutual difference between two source spectra during the supervised training procedure. This discrimination information acts as a kind of regularization so as to maintain between-source separation in monaural source separation. In the experiments, we demonstrate the effectiveness of the proposed method for speech separation compared with the other methods.
引用
收藏
页码:2544 / 2548
页数:5
相关论文
共 50 条
  • [31] Comparative Analysis of Deep Recurrent Neural Networks for Speech Recognition
    Atosha, Pascal Bahavu
    Ozbilge, Emre
    Kirsal, Yonal
    32ND IEEE SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU 2024, 2024,
  • [32] Multitalker Speech Separation With Utterance-Level Permutation Invariant Training of Deep Recurrent Neural Networks
    Kolbaek, Morten
    Yu, Dong
    Tan, Zheng-Hua
    Jensen, Jesper
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2017, 25 (10) : 1901 - 1913
  • [33] Deep Recurrent Neural Networks with Nonlinear Masking Layers and Two-Level Estimation for Speech Separation
    Zhang, Jiantao
    Zhang, Pingjian
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: TEXT AND TIME SERIES, PT IV, 2019, 11730 : 397 - 411
  • [34] PHASE-SENSITIVE AND RECOGNITION-BOOSTED SPEECH SEPARATION USING DEEP RECURRENT NEURAL NETWORKS
    Erdogan, Hakan
    Hershey, John R.
    Watanabe, Shinji
    Le Roux, Jonathan
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 708 - 712
  • [35] DEEP CASA FOR TALKER-INDEPENDENT MONAURAL SPEECH SEPARATION
    Liu, Yuzhou
    Delfarah, Masood
    Wang, DeLiang
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6354 - 6358
  • [36] Binaural reverberant Speech separation based on deep neural networks
    Zhang, Xueliang
    Wang, DeLiang
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 2018 - 2022
  • [37] Speech Separation of A Target Speaker Based on Deep Neural Networks
    Du Jun
    Tu Yanhui
    Xu Yong
    Dai Lirong
    Chin-Hui, Lee
    2014 12TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP), 2014, : 473 - 477
  • [38] Joint Optimization of Modified Ideal Radio Mask and Deep Neural Networks for Monaural Speech Enhancement
    Han, Wei
    Wu, Congming
    Zhang, Xiongwei
    Zhang, Qiye
    Bai, Songting
    2017 IEEE 9TH INTERNATIONAL CONFERENCE ON COMMUNICATION SOFTWARE AND NETWORKS (ICCSN), 2017, : 1070 - 1074
  • [39] FURCAX: END-TO-END MONAURAL SPEECH SEPARATION BASED ON DEEP GATED (DE)CONVOLUTIONAL NEURAL NETWORKS WITH ADVERSARIAL EXAMPLE TRAINING
    Shi, Ziqiang
    Lin, Huibin
    Liu, Liu
    Liu, Rujie
    Hayakawa, Shoji
    Han, Jiqing
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 6985 - 6989
  • [40] Monaural Speech Separation Method Based on Recurrent Attention with Parallel Branches
    Yang, Xue
    Bao, Changchun
    Zhang, Xu
    Chen, Xianhong
    INTERSPEECH 2023, 2023, : 3794 - 3798