Improving RNN-T for Domain Scaling Using Semi-Supervised Training with Neural TTS

被引:5
作者
Deng, Yan [1 ]
Zhao, Rui [2 ]
Meng, Zhong [2 ]
Chen, Xie [2 ]
Liu, Bing [1 ]
Li, Jinyu [2 ]
Gong, Yifan [2 ]
He, Lei [1 ]
机构
[1] Microsoft, Beijing, Peoples R China
[2] Microsoft, Redmond, WA USA
来源
INTERSPEECH 2021 | 2021年
关键词
RNN-T; customization; semi-supervised training; neural TTS; SPEECH;
D O I
10.21437/Interspeech.2021-1017
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Recurrent neural network transducer (RNN-T) has shown to be comparable with conventional hybrid model for speech recognition. However, there is still a challenge in out-of-domain scenarios with context or words different from training data. In this paper, we explore the semi-supervised training which optimizes RNN-T jointly with neural text-to-speech (TTS) to better generalize to new domains using domain-specific text data. We apply the method to two tasks: one with out-of-domain context and the other with significant out-of-vocabulary (OOV) words. The results show that the proposed method significantly improves the recognition accuracy in both tasks, resulting in 61.4% and 53.8% relative word error rate (WER) reductions respectively, from a well-trained RNN-T with 65 thousand hours of training data. We do further study on the semi-supervised training methodology: 1) which modules of RNN-T model to be updated; 2) the impact of using different neural TTS models; 3) the performance of using text with different relevancy to target domain. Finally, we compare several RNN-T customization methods, and conclude that semi-supervised training with neural TTS is comparable and complementary with Internal Language Model Estimation (ILME) or biasing.
引用
收藏
页码:751 / 755
页数:5
相关论文
共 34 条
  • [1] Aleksic P, 2015, 16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, P468
  • [2] Aleksic P, 2015, INT CONF ACOUST SPEE, P5172, DOI 10.1109/ICASSP.2015.7178957
  • [3] [Anonymous], 2019, P INTERSPEECH
  • [4] [Anonymous], 2017, LJ SPEECH DATASET
  • [5] Baskar Murali Karthick, 2019, ARXIV190501152
  • [6] Battenberg E, 2017, 2017 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), P206, DOI 10.1109/ASRU.2017.8268937
  • [7] Bell P., 2021, IEEE OPEN J SINGAL P, V2
  • [8] Chan W, 2016, INT CONF ACOUST SPEE, P4960, DOI 10.1109/ICASSP.2016.7472621
  • [9] Chen M., 2021, 2021 INT C LEARNING
  • [10] Improving Speech Recognition using GAN-based Speech Synthesis and Contrastive Unspoken Text Selection
    Chen, Zhehuai
    Rosenberg, Andrew
    Zhang, Yu
    Wang, Gary
    Ramabhadran, Bhuvana
    Moreno, Pedro J.
    [J]. INTERSPEECH 2020, 2020, : 556 - 560