Revisiting squared-error and cross-entropy functions for training neural network classifiers

被引:170
作者
Kline, DM
Berardi, VL [1 ]
机构
[1] Kent State Univ, Grad Sch Management, Kent, OH 44221 USA
[2] Univ N Carolina, Cameron Sch Business, Wilmington, NC 28401 USA
关键词
D O I
10.1007/s00521-005-0467-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper investigates the efficacy of cross-entropy and square-error objective functions used in training feed-forward neural networks to estimate posterior probabilities. Previous research has found no appreciable difference between neural network classifiers trained using cross-entropy or squared-error. The approach employed here, though, shows cross-entropy has significant, practical advantages over squared-error.
引用
收藏
页码:310 / 318
页数:9
相关论文
共 21 条
[1]  
[Anonymous], NEURAL COMPUT
[2]  
BAUM EB, 1988, NEURAL INFORMATION P, P52
[3]   The effect of misclassification costs on neural network classifiers [J].
Berardi, VL ;
Zhang, GP .
DECISION SCIENCES, 1999, 30 (03) :659-682
[4]  
Bishop C. M., 1996, Neural networks for pattern recognition
[5]  
Bourlard H., 1989, Advances in Neural Information Processing Systems, VI, P502
[6]  
Cybenko G., 1989, Mathematics of Control, Signals, and Systems, V2, P303, DOI 10.1007/BF02551274
[7]  
Duda R. O., 1973, PATTERN CLASSIFICATI
[8]  
FAGARASAN F, 1995, ARTIFICIAL NEURAL NE, P223
[9]  
Hampshire J. B., 1990, ADV NEURAL INFORMATI, V2, P203
[10]   APPROXIMATION CAPABILITIES OF MULTILAYER FEEDFORWARD NETWORKS [J].
HORNIK, K .
NEURAL NETWORKS, 1991, 4 (02) :251-257