Multi-task Learning Deep Neural Networks For Speech Feature Denoising

被引:0
|
作者
Huang, Bin [1 ]
Ke, Dengfeng [2 ]
Zheng, Hao [2 ]
Xu, Bo [2 ]
Xu, Yanyan [1 ]
Su, Kaile [3 ]
机构
[1] Beijing Forestry Univ, Sch Informat Sci & Technol, Beijing, Peoples R China
[2] Chinese Acad Sci, Inst Automat, Beijing, Peoples R China
[3] Griffith Univ, Inst Integrated & Intelligent Syst, Brisbane, Qld, Australia
关键词
multi-task learning; feature denoising; deep neural networks; ENHANCEMENT;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Traditional automatic speech recognition (ASR) systems usually get a sharp performance drop when noise presents in speech. To make a robust ASR, we introduce a new model using the multi-task learning deep neural networks (MTL-DNN) to solve the speech denoising task in feature level. In this model, the networks are initialized by pre-training restricted Boltzmann machines (RBM) and fine-tuned by jointly learning multiple interactive tasks using a shared representation. In multi-task learning, we choose a noisy-clean speech pair fitting task as the primary task and separately explore two constraints as the secondary tasks: phone label and phone cluster. In experiments, the denoised speech is reconstructed by the MTL-DNN using the noisy speech as input and it is respectively evaluated by the DNN-hidden Markov model (HMM) based and the Gaussian Mixture Model (GMM)-HMM based ASR systems. Results show that, using the denoised speech, the word error rate (WER) is respectively reduced by 53.14% and 34.84% compared with baselines. The MTL-DNN model also outperforms the general single-task learning deep neural networks (STL-DNN) model with a performance improvement of 4.93% and 3.88% respectively.
引用
收藏
页码:2464 / 2468
页数:5
相关论文
共 50 条
  • [31] Empirical evaluation of multi-task learning in deep neural networks for natural language processing
    Li, Jianquan
    Liu, Xiaokang
    Yin, Wenpeng
    Yang, Min
    Ma, Liqun
    Jin, Yaohong
    NEURAL COMPUTING & APPLICATIONS, 2021, 33 (09): : 4417 - 4428
  • [32] Multi-Task Deep Neural Networks for Natural Language Understanding
    Liu, Xiaodong
    He, Pengcheng
    Chen, Weizhu
    Gao, Jianfeng
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 4487 - 4496
  • [33] Creating CREATE queries with multi-task deep neural networks
    Diker, S. Nazmi
    Sakar, C. Okan
    KNOWLEDGE-BASED SYSTEMS, 2023, 266
  • [34] MULTI-TASK DEEP NEURAL NETWORK FOR MULTI-LABEL LEARNING
    Huang, Yan
    Wang, Wei
    Wang, Liang
    Tan, Tieniu
    2013 20TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2013), 2013, : 2897 - 2900
  • [35] Convex multi-task feature learning
    Andreas Argyriou
    Theodoros Evgeniou
    Massimiliano Pontil
    Machine Learning, 2008, 73 : 243 - 272
  • [36] JOINT ACOUSTIC MODELING OF TRIPHONES AND TRIGRAPHEMES BY MULTI-TASK LEARNING DEEP NEURAL NETWORKS FOR LOW-RESOURCE SPEECH RECOGNITION
    Chen, Dongpeng
    Mak, Brian
    Leung, Cheung-Chi
    Sivadas, Sunil
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [37] Convex multi-task feature learning
    Argyriou, Andreas
    Evgeniou, Theodoros
    Pontil, Massimiliano
    MACHINE LEARNING, 2008, 73 (03) : 243 - 272
  • [38] Multi-Task Feature Interaction Learning
    Lin, Kaixiang
    Xu, Jianpeng
    Baytas, Inci M.
    Ji, Shuiwang
    Zhou, Jiayu
    KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 1735 - 1744
  • [39] Multi-task Learning Neural Networks for Comparative Elements Extraction
    Liu, Dianqing
    Wang, Lihui
    Shao, Yanqiu
    CHINESE LEXICAL SEMANTICS (CLSW 2020), 2021, 12278 : 398 - 407
  • [40] Deep Elastic Networks with Model Selection for Multi-Task Learning
    Ahn, Chanho
    Kim, Eunwoo
    Oh, Songhwai
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6528 - 6537