The latent words language model

被引:18
作者
Deschacht, Koen [1 ]
De Belder, Jan [1 ]
Moens, Marie-Francine [1 ]
机构
[1] Katholieke Univ Leuven, Dept Comp Sci, B-3001 Heverlee, Belgium
关键词
Language model; Information extraction; Word sense disambiguation; Semantic role labeling; COMPLEXITY;
D O I
10.1016/j.csl.2012.04.001
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a new generative model of natural language, the latent words language model. This model uses a latent variable for every word in a text that represents synonyms or related words in the given context. We develop novel methods to train this model and to find the expected value of these latent variables for a given unseen text. The learned word similarities help to reduce the sparseness problems of traditional n-gram language models. We show that the model significantly outperforms interpolated Kneser-Ney smoothing and class-based language models on three different corpora. Furthermore the latent variables are useful features for information extraction. We show that both for semantic role labeling and word sense disambiguation, the performance of a supervised classifier increases when incorporating these variables as extra features. This improvement is especially large when using only a small annotated corpus for training. (C) 2012 Elsevier Ltd. All rights reserved.
引用
收藏
页码:384 / 409
页数:26
相关论文
共 80 条
  • [1] Ando RK, 2005, J MACH LEARN RES, V6, P1817
  • [2] [Anonymous], COMPUTATIONAL LINGUI
  • [3] [Anonymous], 1992, COLING 1992, DOI DOI 10.3115/992133.992154
  • [4] [Anonymous], 2005, Advances in Neural Information Processing Systems
  • [5] [Anonymous], 2007, P 45 ANN M ASS COMP
  • [6] [Anonymous], P 12 C COMP NAT LANG
  • [7] [Anonymous], COMPUTATIONAL LINGUI
  • [8] [Anonymous], P 2004 HUM LANG TECH
  • [9] [Anonymous], P 5 EUR C SPEECH COM
  • [10] [Anonymous], 1992, AAAI S PROB APPR NAT