Consistency analysis of an empirical minimum error entropy algorithm

被引:49
作者
Fan, Jun [1 ]
Hu, Ting [2 ]
Wu, Qiang [3 ]
Zhou, Ding-Xuan [4 ]
机构
[1] Univ Wisconsin, Dept Stat, 1300 Univ Ave, Madison, WI 53706 USA
[2] Wuhan Univ, Sch Math & Stat, Wuhan 430072, Peoples R China
[3] Middle Tennessee State Univ, Dept Math Sci, Murfreesboro, TN 37132 USA
[4] City Univ Hong Kong, Dept Math, Kowloon, Hong Kong, Peoples R China
关键词
Minimum error entropy; Learning theory; Renyi's entropy; Error entropy consistency; Regression consistency; CRITERION;
D O I
10.1016/j.acha.2014.12.005
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
In this paper we study the consistency of an empirical minimum error entropy (MEE) algorithm in a regression setting. We introduce two types of consistency. The error entropy consistency, which requires the error entropy of the learned function to approximate the minimum error entropy, is shown to be always true if the bandwidth parameter tends to 0 at an appropriate rate. The regression consistency, which requires the learned function to approximate the regression function, however, is a complicated issue. We prove that the error entropy consistency implies the regression consistency for homoskedastic models where the noise is independent of the input variable. But for heteroskedastic models, a counterexample is used to show that the two types of consistency do not coincide. A surprising result is that the regression consistency is always true, provided that the bandwidth parameter tends to infinity at an appropriate rate. Regression consistency of two classes of special models is shown to hold with fixed bandwidth parameter, which further illustrates the complexity of regression consistency of MEE. Fourier transform plays crucial roles in our analysis. Published by Elsevier Inc.
引用
收藏
页码:164 / 189
页数:26
相关论文
共 19 条
[1]  
[Anonymous], 1961, P AM MATH SOC
[2]  
[Anonymous], 1999, EXTREMES
[3]  
Bartlett P. L., 2003, Journal of Machine Learning Research, V3, P463, DOI 10.1162/153244303321897690
[4]   Local Rademacher complexities [J].
Bartlett, PL ;
Bousquet, O ;
Mendelson, S .
ANNALS OF STATISTICS, 2005, 33 (04) :1497-1537
[5]   Convergence properties and data efficiency of the minimum error entropy criterion in adaline training [J].
Erdogmus, D ;
Principe, JC .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2003, 51 (07) :1966-1978
[6]   Blind source separation using Renyi's α-marginal entropies [J].
Erdogmus, D ;
Hild, KE ;
Principe, JC .
NEUROCOMPUTING, 2002, 49 :25-38
[7]   An error-entropy minimization algorithm for supervised training of nonlinear adaptive systems [J].
Erdogmus, D ;
Principe, JC .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2002, 50 (07) :1780-1786
[8]  
ERDOGMUS D, 2000, P 2 INT WORKSH IND C, P75
[9]   Information theoretic clustering [J].
Gokcay, E ;
Principe, JC .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2002, 24 (02) :158-171
[10]  
Hu T, 2013, J MACH LEARN RES, V14, P377