Scalable Generalized Dynamic Topic Models

被引:0
作者
Jaehnichen, Patrick [1 ]
Wenzel, Florian [1 ,2 ]
Kloft, Marius [1 ,2 ]
Mandt, Stephan [3 ]
机构
[1] Humboldt Univ, Berlin, Germany
[2] TU Kaiserslautern, Kaiserslautern, Germany
[3] Disney Res, Los Angeles, CA USA
来源
INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84 | 2018年 / 84卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dynamic topic models (DTMs) model the evolution of prevalent themes in literature, online media, and other forms of text over time. DTMs assume that word co-occurrence statistics change continuously and therefore impose continuous stochastic process priors on their model parameters. These dynamical priors make inference much harder than in regular topic models, and also limit scalability. In this paper, we present several new results around DTMs. First, we extend the class of tractable priors from Wiener processes to the generic class of Gaussian processes (GPs). This allows us to explore topics that develop smoothly over time, that have a long-term memory or are temporally concentrated (for event detection). Second, we show how to perform scalable approximate inference in these models based on ideas around stochastic variational inference and sparse Gaussian processes. This way we can train a rich family of DTMs to massive data. Our experiments on several large-scale datasets show that our generalized model allows us to find interesting patterns that were not accessible by previous approaches.
引用
收藏
页数:9
相关论文
共 36 条
  • [1] Agovic A., 2012, ARXIV12033462
  • [2] Amari S., 2007, METHODS INFORM GEOME
  • [3] [Anonymous], 2017, ARXIV170208359
  • [4] [Anonymous], 2005, P CVPR
  • [5] [Anonymous], P ICML
  • [6] [Anonymous], P AISTATS
  • [7] Bhadury A., 2016, P WWW
  • [8] Probabilistic Topic Models
    Blei, David M.
    [J]. COMMUNICATIONS OF THE ACM, 2012, 55 (04) : 77 - 84
  • [9] Blei David M, 2007, ANN APPL STAT
  • [10] Latent Dirichlet allocation
    Blei, DM
    Ng, AY
    Jordan, MI
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) : 993 - 1022