Dynamic Embeddings for Language Evolution

被引:79
作者
Rudolph, Maja [1 ]
Blei, David [1 ]
机构
[1] Columbia Univ, New York, NY 10027 USA
来源
WEB CONFERENCE 2018: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW2018) | 2018年
关键词
word embeddings; exponential family embeddings; probabilistic modeling; dynamic modeling; semantic change;
D O I
10.1145/3178876.3185999
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Word embeddings are a powerful approach for unsupervised analysis of language. Recently, Rudolph et al. [35] developed exponential family embeddings, which cast word embeddings in a probabilistic framework. Here, we develop dynamic embeddings, building on exponential family embeddings to capture how the meanings of words change over time. We use dynamic embeddings to analyze three large collections of historical texts: the U.S. Senate speeches from 1858 to 2009, the history of computer science ACM abstracts from 1951 to 2014, and machine learning papers on the ArXiv from 2007 to 2015. We find dynamic embeddings provide better fits than classical embeddings and capture interesting patterns about how language changes.
引用
收藏
页码:1003 / 1011
页数:9
相关论文
共 48 条
[1]  
Aitchison J., 2001, LANGUAGE CHANGE PROG, V3
[2]  
[Anonymous], 2017, CORR
[3]  
[Anonymous], 2017, ARXIV170208359
[4]  
[Anonymous], 2014, LTCSS ACL, DOI DOI 10.3115/V1/W14-2517
[5]  
[Anonymous], 2009, Advances in neural information processing systems
[6]  
[Anonymous], 2014, Transactions of the Association for Computational Linguistics
[7]  
[Anonymous], 2005, INT WORKSHOP ARTIFIC
[8]  
[Anonymous], 2013, Advances in Neural Information Processing Systems
[9]  
[Anonymous], 2016, Edward: A library for probabilistic modeling, inference, and criticism
[10]  
[Anonymous], 2016, Advances in Neural Information Processing Systems