CREATING ARTIFICIAL NEURAL NETWORKS THAT GENERALIZE

被引:346
作者
SIETSMA, J
DOW, RJF
机构
关键词
NEURAL NETWORKS; BACK-PROPAGATION; PATTERN RECOGNITION; GENERALIZATION; HIDDEN UNITS; PRUNING;
D O I
10.1016/0893-6080(91)90033-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We develop a technique to test the hypothesis that multilayered, feed-forward networks with few units on the first hidden layer generalize better than networks with many units in the first layer. Large networks are trained to perform a classification task and the redundant units are removed ("pruning") to produce the smallest network capable of performing the task. A technique for inserting layers where pruning has introduced linear inseparability is also described. Two tests of ability to generalize are used - the ability to classify training inputs corrupted by noise and the ability to classify new patterns from each class. The hypothesis is found to be false for networks trained with noisy inputs. Pruning to the minimum number of units in the first layer produces networks which correctly classify the training set but generalize poorly compared with larger networks.
引用
收藏
页码:67 / 79
页数:13
相关论文
共 13 条