Boosted Mixture Learning of Gaussian Mixture HMMs for Speech Recognition

被引:0
作者
Du, Jun [1 ]
Hu, Yu [1 ]
Jiang, Hui [2 ]
机构
[1] iFlytek Res, Hefei, Anhui, Peoples R China
[2] York Univ, Toronto, ON M3J 1P3, Canada
来源
11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 3 AND 4 | 2010年
关键词
functional gradient; boosted mixture learning; acoustic models; speech recognition; MODEL;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this paper, we propose a novel boosted mixture learning (BML) framework for Gaussian mixture HMMs in speech recognition. BML is an incremental method to learn mixture models for classification problem. In each step of BML, one new mixture component is calculated according to functional gradient of an objective function to ensure that it is added along the direction to maximize the objective function the most. Several techniques have been proposed to extend BML from simple mixture models like Gaussian mixture model (GMM) to Gaussian mixture hidden Markov model (HMM), including Viterbi approximation to obtain state segmentation, weight decay to initialize sample weights to avoid overfitting, combining partial updating with global updating of parameters and using Bayesian information criterion (BIC) for parsimonious modeling. Experimental results on the WSJ0 task have shown that the proposed BML yields relative word and sentence error rate reduction of 10.9% and 12.9%, respectively, over the conventional training procedure.
引用
收藏
页码:2942 / +
页数:2
相关论文
共 7 条
  • [1] Kim M., 2007, P ICML, P409
  • [2] Mason L., 1999, NIPS
  • [3] MCLACHLAN G., 2001, FINITE MIXTURE MODEL
  • [4] Pavlovic V, 2004, PROC CVPR IEEE, P811
  • [5] ROSSET S, 2005, P 11 ACM SIGKDD INT, P249
  • [6] ESTIMATING DIMENSION OF A MODEL
    SCHWARZ, G
    [J]. ANNALS OF STATISTICS, 1978, 6 (02) : 461 - 464
  • [7] WOODLAND PC, 1994, INT CONF ACOUST SPEE, P125