Distant bigram language modelling using maximum entropy

被引:0
|
作者
Simons, M
Ney, H
Martin, SC
机构
来源
1997 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I - V: VOL I: PLENARY, EXPERT SUMMARIES, SPECIAL, AUDIO, UNDERWATER ACOUSTICS, VLSI; VOL II: SPEECH PROCESSING; VOL III: SPEECH PROCESSING, DIGITAL SIGNAL PROCESSING; VOL IV: MULTIDIMENSIONAL SIGNAL PROCESSING, NEURAL NETWORKS - VOL V: STATISTICAL SIGNAL AND ARRAY PROCESSING, APPLICATIONS | 1997年
关键词
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper, we apply tile maximum entropy approach to so-called distant bigram language modelling. In addition to the usual unigram and bigram dependencies, we use distant bigram dependencies, where tile immediate predecessor word of the word position under consideration is skipped. The contributions of this paper are: (1) We analyze the computational complexity of the resulting training algorithm, i.e. the generalized iterative scaling (GIS) algorithm, and studs the details of its implementation. (2) We describe a method for handling unseen events in the maximum entropy approach; this is achieved by discounting the frequencies of observed events. (3) We study the effect of this discounting operation on the convergence of the GIS algorithm. (4) We give experimental perplexity results for a corpus from the WSJ task. By using the maximum entropy approach and the distant bigram dependencies, we are able to reduce the perplexity from 205.4 for our best conventional bigram model to 169.5.
引用
收藏
页码:787 / 790
页数:4
相关论文
共 50 条
  • [31] Maximum entropy snapshot sampling for reduced basis modelling
    Bannenberg, Marcus W. F. M.
    Kasolis, Fotios
    Guenther, Michael
    Clemens, Markus
    COMPEL-THE INTERNATIONAL JOURNAL FOR COMPUTATION AND MATHEMATICS IN ELECTRICAL AND ELECTRONIC ENGINEERING, 2022, 41 (03) : 954 - 966
  • [32] Incorporating linguistic structure into maximum entropy language models
    GaoLin Fang
    Wen Gao
    ZhaoQi Wang
    Journal of Computer Science and Technology, 2003, 18 : 131 - 136
  • [33] Latent maximum entropy principle for statistical language modeling
    Wang, SJ
    Rosenfeld, R
    Zhao, YX
    ASRU 2001: IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, CONFERENCE PROCEEDINGS, 2001, : 182 - 185
  • [34] Improved maximum entropy language model and its application
    Li, Juanzi
    Huang, Changning
    Ruan Jian Xue Bao/Journal of Software, 1999, 10 (03): : 257 - 263
  • [35] Learning to parse natural language with maximum entropy models
    Ratnaparkhi, A
    MACHINE LEARNING, 1999, 34 (1-3) : 151 - 175
  • [36] Digging Language Model - Maximum Entropy Phrase Extraction
    Kanis, Jakub
    TEXT, SPEECH, AND DIALOGUE, 2016, 9924 : 46 - 53
  • [37] Learning to Parse Natural Language with Maximum Entropy Models
    Adwait Ratnaparkhi
    Machine Learning, 1999, 34 : 151 - 175
  • [38] Incorporating linguistic structure into maximum entropy language models
    Fang, GL
    Gao, W
    Wang, ZQ
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2003, 18 (01) : 131 - 136
  • [39] Maximum Entropy Named Entity Recognition for Czech Language
    Konkol, Michal
    Konopik, Miloslav
    TEXT, SPEECH AND DIALOGUE, TSD 2011, 2011, 6836 : 203 - 210
  • [40] Modulus maximum image energy using maximum entropy
    Heric, D
    Zazula, D
    PROCEEDINGS ELMAR-2005, 2005, : 57 - 60