Advanced search algorithms for information-theoretic learning with kernel-based estimators

被引:30
作者
Morejon, RA [1 ]
Principe, JC [1 ]
机构
[1] Univ Florida, Computat Neuroengn Lab, Gainesville, FL 32611 USA
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2004年 / 15卷 / 04期
基金
美国国家科学基金会;
关键词
adaptive parameters; entropy; kernel estimation; mutual information; training algorithms;
D O I
10.1109/TNN.2004.828769
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent publications have proposed various information-theoretic learning (ITL) criteria based on Renyi's quadratic entropy with nonparametric kernel-based density estimation as alternative performance metrics for both supervised and unsupervised adaptive system training. These metrics, based on entropy and mutual information, take into account higher order statistics unlike the mean-square error (MSE) criterion. The drawback of these information-based metrics is the increased computational complexity, which underscores the importance of efficient training algorithms. In this paper, we examine familiar advanced-parameter search algorithms and propose modifications to allow training of systems with these ITL criteria. The well known algorithms tailored here for ITL include various improved gradient-descent methods, conjugate gradient approaches, and the Levenberg-Marquardt (LM) algorithm. Sample problems and metrics are presented to illustrate the computational efficiency attained by employing the proposed algorithms.
引用
收藏
页码:874 / 884
页数:11
相关论文
共 17 条
[1]   1ST-ORDER AND 2ND-ORDER METHODS FOR LEARNING - BETWEEN STEEPEST DESCENT AND NEWTON METHOD [J].
BATTITI, R .
NEURAL COMPUTATION, 1992, 4 (02) :141-166
[2]  
Bishop C. M., 1996, Neural networks for pattern recognition
[3]  
Demuth H., 2004, Neural Network Toolbox For Use with MATLAB (Version 4)
[4]   Online entropy manipulation: Stochastic information gradient [J].
Erdogmus, D ;
Hild, KE ;
Principe, JC .
IEEE SIGNAL PROCESSING LETTERS, 2003, 10 (08) :242-245
[5]   Generalized information potential criterion for adaptive system training [J].
Erdogmus, D ;
Principe, JC .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2002, 13 (05) :1035-1044
[6]   An error-entropy minimization algorithm for supervised training of nonlinear adaptive systems [J].
Erdogmus, D ;
Principe, JC .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2002, 50 (07) :1780-1786
[7]  
ERDOGMUS D, 2002, P 11 EUR SIGN PROC C
[8]  
ERDOGMUS D, 2000, P 2 INT WORKSH IND C, P75
[9]  
Hagan MT., 1996, NEURAL NETWORK DESIG
[10]   Blind source separation using Renyi's mutual information [J].
Hild, KE ;
Erdogmus, D ;
Príncipe, J .
IEEE SIGNAL PROCESSING LETTERS, 2001, 8 (06) :174-176