Personalizing Recurrent-Neural-Network-Based Language Model by Social Network

被引:19
作者
Lee, Hung-Yi [1 ]
Tseng, Bo-Hsiang [2 ]
Wen, Tsung-Hsien [3 ]
Tsao, Yu
机构
[1] Natl Taiwan Univ, Dept Elect Engn, Taipei 10617, Taiwan
[2] Natl Taiwan Univ, Taipei 10617, Taiwan
[3] Univ Cambridge, Dialogue Syst Grp, Cambridge CB2 1TN, England
关键词
Personalized language modeling; recurrent neural network; social network; ADAPTATION; SYSTEM;
D O I
10.1109/TASLP.2016.2635445
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
With the popularity of mobile devices, personalized speech recognizers have become more attainable and are highly attractive. Since each mobile device is used primarily by a single user, it is possible to have a personalized recognizer that well matches the characteristics of the individual user. Although acoustic model personalization has been investigated for decades, much less work has been reported on personalizing language models, presumably because of the difficulties in collecting sufficient personalized corpora. In this paper, we propose a general framework for personalizing recurrent-neural-network-based language models (RNNLMs) using data collected from social networks, including the posts of many individual users and friend relationships among the users. Two major directions for this are model-based and feature-based RNNLM personalization. In model-based RNNLM personalization, the RNNLM parameters are fine-tuned to an individual user's wording patterns by incorporating social texts posted by the target user and his or her friends. For the feature-based approach, the RNNLM model parameters are fixed across users, but the RNNLM input features are instead augmented with personalized information. Both approaches not only drastically reduce the model perplexity, but also moderately reduce word error rates in n-best rescoring tests.
引用
收藏
页码:519 / 530
页数:12
相关论文
共 63 条
[1]  
[Anonymous], 2011, Rnnlm-recurrent neural network language modeling toolkit
[2]  
[Anonymous], 2012, INTERSPEECH 2012
[3]  
[Anonymous], 2012, P NAACL HLT 2012 WOR
[4]   Statistical language model adaptation: review and perspectives [J].
Bellegarda, JR .
SPEECH COMMUNICATION, 2004, 42 (01) :93-108
[5]   A neural probabilistic language model [J].
Bengio, Y ;
Ducharme, R ;
Vincent, P ;
Jauvin, C .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (06) :1137-1155
[6]  
Berberich K, 2010, LECT NOTES COMPUT SC, V5993, P13, DOI 10.1007/978-3-642-12275-0_5
[7]   Latent Dirichlet allocation [J].
Blei, DM ;
Ng, AY ;
Jordan, MI .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) :993-1022
[8]  
Brown P. F., 1992, Computational Linguistics, V18, P467
[9]   Structured language modeling [J].
Chelba, C ;
Jelinek, F .
COMPUTER SPEECH AND LANGUAGE, 2000, 14 (04) :283-332
[10]   Efficient Training and Evaluation of Recurrent Neural Network Language Models for Automatic Speech Recognition [J].
Chen, Xie ;
Liu, Xunying ;
Wang, Yongqiang ;
Gales, Mark J. F. ;
Woodland, Philip C. .
IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2016, 24 (11) :2146-2157