USING KL-DIVERGENCE AND MULTILINGUAL INFORMATION TO IMPROVE ASR FOR UNDER-RESOURCED LANGUAGES
被引:0
作者:
Imseng, David
论文数: 0引用数: 0
h-index: 0
机构:
Idiap Res Inst, Martigny, SwitzerlandIdiap Res Inst, Martigny, Switzerland
Imseng, David
[1
]
Bourlard, Herve
论文数: 0引用数: 0
h-index: 0
机构:
Idiap Res Inst, Martigny, SwitzerlandIdiap Res Inst, Martigny, Switzerland
Bourlard, Herve
[1
]
Garner, Philip N.
论文数: 0引用数: 0
h-index: 0
机构:
Idiap Res Inst, Martigny, SwitzerlandIdiap Res Inst, Martigny, Switzerland
Garner, Philip N.
[1
]
机构:
[1] Idiap Res Inst, Martigny, Switzerland
来源:
2012 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP)
|
2012年
关键词:
Multilingual speech recognition;
neural network features;
fast training;
Kullback-Leibler divergence;
D O I:
暂无
中图分类号:
O42 [声学];
学科分类号:
070206 ;
082403 ;
摘要:
Setting out from the point of view that automatic speech recognition (ASR) ought to benefit from data in languages other than the target language, we propose a novel Kullback-Leibler (KL) divergence based method that is able to exploit multilingual information in the form of universal phoneme posterior probabilities conditioned on the acoustics. We formulate a means to train a recognizer on several different languages, and subsequently recognize speech in a target language for which only a small amount of data is available. Taking the Greek SpeechDat(II) data as an example, we show that the proposed formulation is sound, and show that it is able to outperform a current state-of-the-art HMM/GMM system. We also use a hybrid Tandem-like system to further understand the source of the benefit.