Memorized Variational Continual Learning for Dirichlet Process Mixtures

被引:0
作者
Yang, Yang [1 ]
Chen, Bo [1 ,2 ]
Liu, Hongwei [1 ,2 ]
机构
[1] Xidian Univ, Natl Lab Radar Signal Proc, Xian 710071, Shaanxi, Peoples R China
[2] Xidian Univ, Collaborat Innovat Ctr Informat Sensing & Underst, Xian 710071, Shaanxi, Peoples R China
关键词
Data models; Task analysis; Bayes methods; Mixture models; Computational modeling; Inference algorithms; Approximation algorithms; Bayesian nonparametric; streaming data; variational continual learning; Dirichlet process mixture; memorized sufficient statistics; discrete and real-valued datasets; INFERENCE;
D O I
10.1109/ACCESS.2019.2947722
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Bayesian nonparametric models are theoretically suitable for streaming data due to their ability to adapt model complexity with the observed data. However, very limited work has addressed posterior inference in a streaming fashion, and most of the existing variational inference algorithms require truncation on variational distributions which cannot vary with the data. In this paper, we focus Dirichlet process mixture models and develop the corresponding variational continual learning approach by maintaining memorized sufficient statistics for previous tasks, called memorized variational continual learning (MVCL), which is able to handle both the posterior update and data in a continual learning setting. Furthermore, we extend MVCL for two cases of mixture models which can handle different data types. The experiments demonstrate the comparable inference capability of our MVCL for both discrete and real-valued datasets with automatically inferring the number of mixture components.
引用
收藏
页码:150851 / 150862
页数:12
相关论文
共 34 条
[1]  
[Anonymous], 2013, NIPS 13, DOI DOI 10.5555/2999611.2999738
[2]  
[Anonymous], 2012, P 25 INT C NEUR INF
[3]   Variational Inference for Dirichlet Process Mixtures [J].
Blei, David M. ;
Jordan, Michael I. .
BAYESIAN ANALYSIS, 2006, 1 (01) :121-143
[4]  
Broderick Tamara, 2013, Advances in Neural Information Processing Systems, V26, P1727
[5]  
Bui T., 2016, PR MACH LEARN RES, P1472
[6]   Markov chain Monte Carlo convergence diagnostics: A comparative review [J].
Cowles, MK ;
Carlin, BP .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1996, 91 (434) :883-904
[7]  
Dheeru D., 2017, TECH REP
[8]  
Duda RO., 2012, Pattern classification
[9]   A Survey of Clustering Algorithms for Big Data: Taxonomy and Empirical Analysis [J].
Fahad, Adil ;
Alshatri, Najlaa ;
Tari, Zahir ;
Alamri, Abdullah ;
Khalil, Ibrahim ;
Zomaya, Albert Y. ;
Foufou, Sebti ;
Bouras, Abdelaziz .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2014, 2 (03) :267-279
[10]  
Ghahramani Z., 2000, P ADV NEUR INF PROC