Perceptron: Learning, Generalization, Model Selection, Fault Tolerance, and Role in the Deep Learning Era

被引:39
作者
Du, Ke-Lin [1 ]
Leung, Chi-Sing [2 ]
Mow, Wai Ho [3 ]
Swamy, M. N. S. [1 ]
机构
[1] Concordia Univ, Dept Elect & Comp Engn, Montreal, PQ H3G 1M8, Canada
[2] City Univ Hong Kong, Dept Elect Engn, Hong Kong, Peoples R China
[3] Hong Kong Univ Sci & Technol, Dept Elect & Comp Engn, Hong Kong, Peoples R China
关键词
multilayer perceptron; perceptron; backpropagation; stochastic gradient descent; second-order learning; model selection; robust learning; deep learning; FEEDFORWARD NEURAL-NETWORKS; CONJUGATE-GRADIENT ALGORITHM; EXTENDED KALMAN FILTER; ERROR BACKPROPAGATION ALGORITHM; WEIGHT INITIALIZATION METHOD; FUZZY MEMBERSHIP FUNCTIONS; MULTILAYER PERCEPTRONS; TRAINING ALGORITHM; BACK-PROPAGATION; PRUNING ALGORITHM;
D O I
10.3390/math10244730
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
The single-layer perceptron, introduced by Rosenblatt in 1958, is one of the earliest and simplest neural network models. However, it is incapable of classifying linearly inseparable patterns. A new era of neural network research started in 1986, when the backpropagation (BP) algorithm was rediscovered for training the multilayer perceptron (MLP) model. An MLP with a large number of hidden nodes can function as a universal approximator. To date, the MLP model is the most fundamental and important neural network model. It is also the most investigated neural network model. Even in this AI or deep learning era, the MLP is still among the few most investigated and used neural network models. Numerous new results have been obtained in the past three decades. This survey paper gives a comprehensive and state-of-the-art introduction to the perceptron model, with emphasis on learning, generalization, model selection and fault tolerance. The role of the perceptron model in the deep learning era is also described. This paper provides a concluding survey of perceptron learning, and it covers all the major achievements in the past seven decades. It also serves a tutorial for perceptron learning.
引用
收藏
页数:46
相关论文
共 391 条
[1]   Noise-boosted bidirectional backpropagation and adversarial learning [J].
Adigun, Olaoluwa ;
Kosko, Bart .
NEURAL NETWORKS, 2019, 120 :9-31
[2]   The "weight smoothing" regularization of MLP for Jacobian stabilization [J].
Aires, F ;
Schmitt, M ;
Chedin, A ;
Scott, N .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1999, 10 (06) :1502-1510
[3]   NEW LOOK AT STATISTICAL-MODEL IDENTIFICATION [J].
AKAIKE, H .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1974, AC19 (06) :716-723
[4]   Natural gradient works efficiently in learning [J].
Amari, S .
NEURAL COMPUTATION, 1998, 10 (02) :251-276
[5]  
Amari S, 1996, ADV NEUR IN, V8, P176
[6]   Information geometry of the EM and em algorithms for neural networks [J].
Amari, SI .
NEURAL NETWORKS, 1995, 8 (09) :1379-1408
[7]   Two highly efficient second-order algorithms for training feedforward networks [J].
Ampazis, N ;
Perantonis, SJ .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2002, 13 (05) :1064-1074
[8]  
ANDERSON J A, 1972, Mathematical Biosciences, V14, P197, DOI 10.1016/0025-5564(72)90075-2
[9]  
[Anonymous], 1962, Perceptions and the Theory of Brain Mechanisms, DOI DOI 10.21236/AD0256582
[10]  
[Anonymous], 1994, A comprehensive foundation, DOI [10.1142/S0129065794000372, DOI 10.1142/S0129065794000372]