Asymptotic convergence properties of the EM algorithm with respect to the overlap in the mixture

被引:15
作者
Ma, JW [1 ]
Xu, L
机构
[1] Peking Univ, Sch Math Sci, Dept Informat Sci, Beijing 100871, Peoples R China
[2] Peking Univ, LMAM, Beijing 100871, Peoples R China
[3] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Shatin, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
EM algorithm; maximum likelihood; asymptotic convergence rate; mixture of densities from exponential families; Gaussian mixtures;
D O I
10.1016/j.neucom.2004.12.009
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The EM algorithm is generally considered as a linearly convergent algorithm. However, many empirical results show that it can converge significantly faster than those gradient based first-order iterative algorithms, especially when the overlap of densities in a mixture is small. This paper explores this issue theoretically on mixtures of densities from a class of exponential families. We have proved that as an average overlap measure of densities in the mixture tends to zero, the asymptotic convergence rate of the EM algorithm locally around the true solution is a higher order infinitesimal than a positive order power of this overlap measure. Thus, the large sample local convergence rate for the EM algorithm tends to be asymptotically superlinear when the overlap of densities in the mixture tends to zero. Moreover, this result has been detailed on Gaussian mixtures. (c) 2005 Elsevier B.V. All rights reserved.
引用
收藏
页码:105 / 129
页数:25
相关论文
共 20 条
[1]  
BARNDORFNIELSEN O, 1978, INFORMATION EXPONENT
[2]  
Delyon B, 1999, ANN STAT, V27, P94
[3]   MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM [J].
DEMPSTER, AP ;
LAIRD, NM ;
RUBIN, DB .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01) :1-38
[4]  
HORNG SC, 1987, P STAT COMP SECT AM, P266
[5]   Acceleration of the EM algorithm by using quasi-Newton methods [J].
Jamshidian, M ;
Jennrich, RI .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1997, 59 (03) :569-587
[6]   CONJUGATE-GRADIENT ACCELERATION OF THE EM ALGORITHM [J].
JAMSHIDIAN, M ;
JENNRICH, RI .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1993, 88 (421) :221-228
[7]   Convergence results for the EM approach to mixtures of experts architectures [J].
Jordan, MI ;
Xu, L .
NEURAL NETWORKS, 1995, 8 (09) :1409-1431
[8]   HIERARCHICAL MIXTURES OF EXPERTS AND THE EM ALGORITHM [J].
JORDAN, MI ;
JACOBS, RA .
NEURAL COMPUTATION, 1994, 6 (02) :181-214
[9]   MAXIMUM-LIKELIHOOD COMPUTATIONS WITH REPEATED MEASURES - APPLICATION OF THE EM-ALGORITHM [J].
LAIRD, N ;
LANGE, N ;
STRAM, D .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1987, 82 (397) :97-105
[10]  
LANGE K, 1995, STAT SINICA, V5, P1