Asymmetric hidden Markov models

被引:12
作者
Bueno, Marcos L. P. [1 ]
Hommersom, Arjen [1 ,2 ]
Lucas, Peter J. F. [1 ,3 ]
Linard, Alexis [1 ]
机构
[1] Radboud Univ Nijmegen, Inst Comp & Informat Sci, Nijmegen, Netherlands
[2] Open Univ, Fac Management Sci & Technol, Heerlen, Netherlands
[3] Leiden Univ, Leiden Inst Adv Comp Sci, Leiden, Netherlands
关键词
Hidden Markov models; Bayesian networks; Model selection; Unsupervised learning; Time series; Structure learning; BAYESIAN NETWORKS; INFERENCE;
D O I
10.1016/j.ijar.2017.05.011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In many problems involving multivariate time series, hidden Markov models (HMMs) are often employed for modeling complex behavior over time. HMMs can, however, require large number of states, what can lead to poor problem insight and model overfitting, especially when limited data is available. In this paper, we further investigate the family of asymmetric hidden Markov models (HMM-As), which generalize the emission distributions to arbitrary Bayesian-network distributions, allowing for state -specific graphical structures in the feature space. As a consequence, HMM-As are able to render more compact state spaces, thus from a learning perspective HMM-As can better handle thecomplexityoverfitting trade-off. In this paper, we study representation properties of asymmetric and symmetric HMMs, as well as provide a learning algorithm for HMM-As. We provide empirical results based on simulations for comparing HMM-As with symmetric and other asymmetry-aware models, showing that modeling more general asymmetries can be very effective. We also consider real-world datasets from several domains, aiming to show that multiple graphical structures underlying data can be identified and are able to provide additional problem insight. Although learning HMM-As can be more complex, it is shown that it is feasible in practice due to their ability to maintain compact state spaces, yet more expressive ones. (C) 2017 Elsevier Inc. All rights reserved.
引用
收藏
页码:169 / 191
页数:23
相关论文
共 53 条
[21]   On probabilistic inference by weighted model counting [J].
Chavira, Mark ;
Darwiche, Adnan .
ARTIFICIAL INTELLIGENCE, 2008, 172 (6-7) :772-799
[22]  
Chickering D., 1996, Learning from data: Artificial intelligence and statistics V, P121, DOI 10.1007/978-1-4612-2404-4_12
[23]   Bayesian Network Structure Learning with Integer Programming: Polytopes, Facets and Complexity [J].
Cussens, James ;
Jarvisalo, Matti ;
Korhonen, Janne H. ;
Bartlett, Mark .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2017, 58 :185-229
[24]   Learning Bayesian networks: approaches and issues [J].
Daly, Ronan ;
Shen, Qiang ;
Aitken, Stuart .
KNOWLEDGE ENGINEERING REVIEW, 2011, 26 (02) :99-157
[25]   MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM [J].
DEMPSTER, AP ;
LAIRD, NM ;
RUBIN, DB .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01) :1-38
[26]   Accelerated Profile HMM Searches [J].
Eddy, Sean R. .
PLOS COMPUTATIONAL BIOLOGY, 2011, 7 (10)
[27]   The hierarchical hidden Markov model: Analysis and applications [J].
Fine, S ;
Singer, Y ;
Tishby, N .
MACHINE LEARNING, 1998, 32 (01) :41-62
[28]  
Freitag D, 2000, SEVENTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-2001) / TWELFTH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE (IAAI-2000), P584
[29]  
Friedman N., 1998, Uncertainty in Artificial Intelligence. Proceedings of the Fourteenth Conference (1998), P129
[30]  
Friedman N, 1996, UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, P252