Bayesian learning of n-gram statistical language modeling

被引:0
作者
Bai, Shuanhu [1 ]
Li, Haizhou [1 ]
机构
[1] Inst Infocomm Res, Singapore, Singapore
来源
2006 IEEE International Conference on Acoustics, Speech and Signal Processing, Vols 1-13 | 2006年
关键词
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The n-gram language model adaptation is typically formulated using deleted interpolation under the maximum likelihood estimation framework. This paper proposes a Bayesian learning framework for n-gram statistical language model training and adaptation. By introducing a Dirichlet conjugate prior to the n-gram parameters, we formulate the deleted interpolation under maximum a posterior criterion with a Bayesian learning procedure. We study the Bayesian learning formulation for n-gram and continuous n-gram language models. The experiments on North American News Text corpus have validated the effectiveness of the proposed algorithms.
引用
收藏
页码:1045 / 1048
页数:4
相关论文
共 8 条
[1]  
BACCHIANI M, ICASSP 2003
[2]  
Brown P. F., 1992, Computational Linguistics, V18, P467
[3]  
CHIEN JT, 2005, INTERSPEECH
[4]  
Huo Q, 1997, IEEE T SPEECH AUDI P, V5, P161, DOI 10.1109/89.554778
[5]  
JELINEK F, 1990, READINGS SPEECH RECO
[6]  
MACKAY DJC, 1994, NAT LANG ENG, V1, P1
[7]   A maximum entropy approach to adaptive statistical language modelling [J].
Rosenfeld, R .
COMPUTER SPEECH AND LANGUAGE, 1996, 10 (03) :187-228
[8]  
SAUL L, 1997, EMNLP