Discriminative Language Modeling With Linguistic and Statistically Derived Features

被引:17
作者
Arisoy, Ebru [1 ]
Saraclar, Murat [2 ]
Roark, Brian [3 ]
Shafran, Izhak [3 ]
机构
[1] IBM Corp, TJ Watson Res Ctr, Speech Technol Media & ACCES Dept, Yorktown Hts, NY 10598 USA
[2] Bogazici Univ, Dept Elect & Elect Engn, TR-34342 Bebek, Turkey
[3] Oregon Hlth & Sci Univ, Ctr Spoken Language Understanding, Beaverton, OR 97006 USA
来源
IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING | 2012年 / 20卷 / 02期
基金
美国国家科学基金会;
关键词
Discriminative training; language modeling; morphologically rich languages; speech recognition; SPEECH RECOGNITION; MORPHOLOGICAL DISAMBIGUATION; TEXT;
D O I
10.1109/TASL.2011.2162323
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This paper focuses on integrating linguistically motivated and statistically derived information into language modeling. We use discriminative language models (DLMs) as a complementary approach to the conventional n-gram language models to benefit from discriminatively trained parameter estimates for overlapping features. In our DLM approach, relevant information is encoded as features. Feature weights are discriminatively trained using training examples and used to re-rank the N-best hypotheses of the baseline automatic speech recognition (ASR) system. In addition to presenting a more complete picture of previously proposed feature sets that extract implicit information available at lexical and sub-lexical levels using both linguistic and statistical approaches, this paper attempts to incorporate semantic information in the form of topic sensitive features. We explore linguistic features to incorporate complex morphological and syntactic language characteristics of Turkish, an agglutinative language with rich morphology, into language modeling. We also apply DLMs to our sub-lexical-based ASR system where the vocabulary is composed of sub-lexical units. Obtaining implicit linguistic information from sub-lexical hypotheses is not as straightforward as word hypotheses, so we use statistical methods to derive useful information from sub-lexical units. DLMs with linguistic and statistical features yield significant, 0.8%-1.1% absolute, improvements over our baseline word-based and sub-word-based ASR systems. The explored features can be easily extended to DLM for other languages.
引用
收藏
页码:540 / 550
页数:11
相关论文
共 48 条
[1]  
Alumae T., 2006, THESIS TALLINN U TEC
[2]  
[Anonymous], 2004, ACL
[3]  
[Anonymous], P M ASS COMP LING AC
[4]   A unified language model for large vocabulary continuous speech recognition of Turkish [J].
Arisoy, Ebru ;
Dutagaci, Helin ;
Arslan, Levent M. .
SIGNAL PROCESSING, 2006, 86 (10) :2844-2862
[5]   SYNTACTIC AND SUB-LEXICAL FEATURES FOR TURKISH DISCRIMINATIVE LANGUAGE MODELS [J].
Arisoy, Ebru ;
Saraclar, Murat ;
Roark, Brian ;
Shafran, Izhak .
2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2010, :5538-5541
[6]  
Arisoy E, 2008, INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, P825
[7]   Turkish Broadcast News Transcription and Retrieval [J].
Arisoy, Ebru ;
Can, Dogan ;
Parlak, Siddika ;
Sak, Hasim ;
Saraclar, Murat .
IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2009, 17 (05) :874-883
[8]   Lattice Extension and Vocabulary Adaptation for Turkish LVCSR [J].
Arisoy, Ebru ;
Saraclar, Murat .
IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2009, 17 (01) :163-173
[9]  
Bacchiani M., 2004, P HUM LANG TECHN C H, P21
[10]  
Bilmes JeffA., 2003, P C HUMAN LANGUAGE T, P4, DOI DOI 10.3115/1073483.1073485