Infinite Switching Dynamic Probabilistic Network With Bayesian Nonparametric Learning

被引:2
作者
Chen, Wenchao [1 ]
Chen, Bo [1 ]
Liu, Yicheng [1 ]
Wang, Chaojie [1 ]
Peng, Xiaojun [2 ]
Liu, Hongwei [1 ]
Zhou, Mingyuan [3 ]
机构
[1] Xidian Univ, Natl Lab Radar Signal Proc, Xian 710071, Peoples R China
[2] Res Acad Rocket, Beijing 100000, Peoples R China
[3] Univ Texas Austin, McCombs Sch Business, Austin, TX 78712 USA
基金
美国国家科学基金会;
关键词
Switches; Dynamical systems; Hidden Markov models; Task analysis; Probabilistic logic; Bayes methods; Data models; Probabilistic dynamical model; dirichlet process mixture; Bayesian nonparametric; SG-MCMC; feature extraction; attention mechanism; MODELS;
D O I
10.1109/TSP.2022.3160535
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
To model sequentially observed multivariate nonstationary count data, we propose a switching Poisson-gamma dynamical systems (SPGDS), a dynamic probabilistic network with switching mechanism. Different from previous models, SPGDS assigns its latent variables into mixture of gamma distributed parameters to model complex sequences and describe the nonlinear dynamics, meanwhile, capture various temporal dependencies. Moreover, SPGDS can model all discrete and nonnegative real data by linking them to latent counts. To take advantage of Bayesian nonparametrics in handling the unknown number of mixture components, we integrate Dirichlet process (DP) mixture into SPGDS and develop an infinite switching Poisson-gamma dynamical systems (iSPGDS). For efficient and nonparametric inference, we develop a infinite switching recurrent variational inference network, combined with a scalable hybrid stochastic gradient-MCMC and variational inference method, which is scalable to large scale sequences and fast in out-of-sample prediction. Besides, to handle the time-series categorization task, we further propose an supervised attention iSPGDS (attn-iSPGDS), which combines the representation power of iSPGDS, discriminative power of deep neural networks, and selection power of the attention mechanism under a principled probabilistic framework. Experiments on both unsupervised and supervised tasks demonstrate that the proposed model not only has excellent fitting and prediction performance on complex sequences, but also separates different dynamical patterns within them.
引用
收藏
页码:2224 / 2238
页数:15
相关论文
共 64 条
[1]  
Aaron S., 2016, Advances in Neural Information Processing Systems, P5006
[2]  
Acharya S, 2015, PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE OF THE ASSOCIATION OF PSYCHOLOGY AND PSYCHIATRY FOR ADULTS AND CHILDREN (A.P.P.A.C 2013)
[3]  
[Anonymous], 2014, P SSST8 8 WORKSH SYN
[4]  
[Anonymous], 2011, Proceedings ICML
[5]  
Arjovsky M, 2016, PR MACH LEARN RES, V48
[6]   Gaussian Mixture Models Reduction by Variational Maximum Mutual Information [J].
Bar-Yosef, Yossi ;
Bistritz, Yuval .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2015, 63 (06) :1557-1569
[7]  
Barone AVM, 2018, DEEP LEARNING APPROACHES FOR LOW-RESOURCE NATURAL LANGUAGE PROCESSING (DEEPLO), P77
[8]  
Becker-Ehmck P., 2019, P 36 INT C MACHINE L, P553
[9]   Variational Inference for Dirichlet Process Mixtures [J].
Blei, David M. ;
Jordan, Michael I. .
BAYESIAN ANALYSIS, 2006, 1 (01) :121-143
[10]  
Campos V., 2018, PROC 6 INT C LEARN R