Analysis of Low-Resource Acoustic Model Self-Training

被引:0
|
作者
Novotney, Scott [1 ]
Schwartz, Richard [1 ]
机构
[1] BBN Technol, Cambridge, MA USA
来源
INTERSPEECH 2009: 10TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2009, VOLS 1-5 | 2009年
关键词
speech recognition; self-training; English;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Previous work on self-training of acoustic models using unlabeled data reported significant reductions in WER assuming a large phonetic dictionary was available. We now assume only those words from ten hours of speech are initially available. Subsequently, we are then given a large vocabulary and then quantify the value of repeating self-training with this larger dictionary. This experiment is used to analyze the effects of self-training on categories of words. We report the following findings: (i) Although the small 5k vocabulary raises WER by 2% absolute, self-training is equally effective as using a large 75k vocabulary. (ii) Adding all 75k words to the decoding vocabulary after self-training reduces the WER degradation to only 0.8% absolute. (iii) Self-training most benefits those words in the unlabeled audio but not transcribed by a wide margin.
引用
收藏
页码:236 / 239
页数:4
相关论文
共 50 条
  • [1] A Novel Self-training Approach for Low-resource Speech Recognition
    Singh, Satwinder
    Hou, Feng
    Wang, Ruili
    INTERSPEECH 2023, 2023, : 1588 - 1592
  • [2] Low-Resource Mandarin Prosodic Structure Prediction Using Self-Training
    Wang, Xingrui
    Zhang, Bowen
    Shinozaki, Takahiro
    2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 859 - 863
  • [3] Self-Training With Double Selectors for Low-Resource Named Entity Recognition
    Fu, Yingwen
    Lin, Nankai
    Yu, Xiaohui
    Jiang, Shengyi
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 1265 - 1275
  • [4] Uncertainty-Aware Self-training for Low-Resource Neural Sequence Labeling
    Wang, Jianing
    Wang, Chengyu
    Huang, Jun
    Gao, Ming
    Zhou, Aoying
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 13682 - 13690
  • [5] Arabic Emotion Recognition in Low-Resource Settings: A Novel Diverse Model Stacking Ensemble with Self-Training
    Althobaiti, Maha Jarallah
    APPLIED SCIENCES-BASEL, 2023, 13 (23):
  • [6] Acoustic model training using self-attention for low-resource speech recognition
    Park, Hosung
    Kim, Ji-Hwan
    JOURNAL OF THE ACOUSTICAL SOCIETY OF KOREA, 2020, 39 (05): : 483 - 489
  • [7] Cross-Lingual Summarization Method Based on Joint Training and Self-Training in Low-Resource Scenarios
    Cheng, Shaohuan
    Tang, Yujia
    Liu, Qiao
    Chen, Wenyu
    Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China, 2024, 53 (05): : 762 - 770
  • [8] Cross-Lingual Self-training to Learn Multilingual Representation for Low-Resource Speech Recognition
    Zi-Qiang Zhang
    Yan Song
    Ming-Hui Wu
    Xin Fang
    Ian McLoughlin
    Li-Rong Dai
    Circuits, Systems, and Signal Processing, 2022, 41 : 6827 - 6843
  • [9] State Value Generation with Prompt Learning and Self-Training for Low-Resource Dialogue State Tracking
    Gu, Ming
    Yang, Yan
    Chen, Chengcai
    Yu, Zhou
    ASIAN CONFERENCE ON MACHINE LEARNING, VOL 222, 2023, 222
  • [10] Cross-Lingual Self-training to Learn Multilingual Representation for Low-Resource Speech Recognition
    Zhang, Zi-Qiang
    Song, Yan
    Wu, Ming-Hui
    Fang, Xin
    McLoughlin, Ian
    Dai, Li-Rong
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2022, 41 (12) : 6827 - 6843