Bag-Of-Word normalized n-gram models

被引:0
作者
Sethy, Abhinav [1 ]
Ramabhadran, Bhuvana [1 ]
机构
[1] IBM TJ Watson Res Ctr, Yorktown Hts, NY USA
来源
INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5 | 2008年
关键词
language modeling; bag-of-words; topic modeling;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Bag-Of-Word (BOW) model uses a fixed length vector of word counts to represent text. Although the model disregards word sequence information, it has been shown to be successful in capturing long range word-word correlations and topic information. In contrast, n-gram models have been shown to be an effective way to capture short term dependencies by modeling text as a Markovian sequence. In this paper, we propose a probabilistic framework to combine BOW models with n-gram models. In the proposed framework, we normalize the n-gram model to build a model for word sequences given the corresponding bag-of-words representation. By combining the two models, the proposed approach allows us to capture the latent topic information as well as local Markovian dependencies in text. Using the proposed model, we were able to achieve a 10% reduction in perplexity and a 2% reduction in WER (relative) over a state-of-the-art baseline for transcribing broadcast news in English.
引用
收藏
页码:1594 / 1597
页数:4
相关论文
共 19 条
[1]  
Bell J. S., 2004, SPEAKABLE UNSPEAKABL, V2nd, P93
[2]   Latent Dirichlet allocation [J].
Blei, DM ;
Ng, AY ;
Jordan, MI .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) :993-1022
[3]   Advances in speech transcription at IBM under the DARPA EARS program [J].
Chen, Stanley F. ;
Kingsbury, Brian ;
Mangu, Lidia ;
Povey, Daniel ;
Saon, George ;
Soltau, Hagen ;
Zweig, Geoffrey .
IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2006, 14 (05) :1596-1608
[4]  
DeerWester S., 1990, J AM SOC INFORM SCI
[5]  
Griffiths T., 2005, NIPS
[6]  
GRIFFITHS TL, 2007, ROAD MEANING
[7]  
HEIDEL A, 2007, P ASRU
[8]  
Hoffman T., 1999, P UNC ART INT
[9]  
Hsu B. J., 2006, P EMNLP
[10]  
Kawahara T., 2008, P ICASSP