Approximation by neural networks with weights varying on a finite set of directions

被引:20
作者
Ismailov, Vugar E. [1 ]
机构
[1] Natl Acad Sci, Inst Math & Mech, AZ-1141 Baku, Azerbaijan
关键词
Neural network; MLP model; Activation function; Weight; Density; Approximation; Path; Orbit;
D O I
10.1016/j.jmaa.2011.11.037
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Approximation properties of the MLP (multilayer feedforward perceptron) model of neural networks have been investigated in a great deal of works over the last 30 years. It has been shown that for a large class of activation functions, a neural network can approximate arbitrarily well any given continuous function. The most significant result on this problem belongs to Leshno, Lin, Pinkus and Schocken. They proved that the necessary and sufficient condition for any single hidden layer network to have the u.a.p. (universal approximation property) is that its activation function not be a polynomial. Some authors (White, Stinchcombe, Ito, and others) showed that a single hidden layer perceptron with some bounded weights can also have the u.a.p. Thus the weights required for u.a.p. are not necessary to be of an arbitrarily, large magnitude. But what if they are too restricted? How can one learn approximation properties of networks with arbitrarily restricted set of weights? The current paper makes a first step in solving this general problem. We consider neural networks with sets of weights consisting of a finite number of directions. Our purpose is to characterize compact sets X in the d-dimensional space such that the network can approximate any continuous function over X. In a special case, when weights vary only on two directions, we give a lower bound for the approximation error and find a sufficient condition for the network to be a best approximation. (C) 2011 Elsevier Inc. All rights reserved.
引用
收藏
页码:72 / 83
页数:12
相关论文
共 25 条
[1]  
[Anonymous], TRANSL MATH MONOGR
[2]   INTERPOLATION BY RIDGE FUNCTIONS [J].
BRAESS, D ;
PINKUS, A .
JOURNAL OF APPROXIMATION THEORY, 1993, 73 (02) :218-236
[3]   APPROXIMATIONS OF CONTINUOUS FUNCTIONALS BY NEURAL NETWORKS WITH APPLICATION TO DYNAMIC-SYSTEMS [J].
CHEN, TP ;
CHEN, H .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1993, 4 (06) :910-918
[4]   APPROXIMATION BY RIDGE FUNCTIONS AND NEURAL NETWORKS WITH ONE HIDDEN LAYER [J].
CHUI, CK ;
LI, X .
JOURNAL OF APPROXIMATION THEORY, 1992, 70 (02) :131-141
[5]  
Cybenko G., 1989, Mathematics of Control, Signals, and Systems, V2, P303, DOI 10.1007/BF02551274
[6]  
DeVore R. A., 1997, Annals of Numerical Mathematics, V4, P261
[7]   Approximation by neural networks with a bounded number of nodes at each level [J].
Gripenberg, G .
JOURNAL OF APPROXIMATION THEORY, 2003, 122 (02) :260-266
[8]   APPROXIMATION CAPABILITIES OF MULTILAYER FEEDFORWARD NETWORKS [J].
HORNIK, K .
NEURAL NETWORKS, 1991, 4 (02) :251-257
[9]   On the representation by linear superpositions [J].
Ismailov, Vugar E. .
JOURNAL OF APPROXIMATION THEORY, 2008, 151 (02) :113-125
[10]   On the theorem of M Golomb [J].
Ismailov, Vugar E. .
PROCEEDINGS OF THE INDIAN ACADEMY OF SCIENCES-MATHEMATICAL SCIENCES, 2009, 119 (01) :45-52