Constructive training methods for feedforward neural networks with binary weights

被引:5
作者
Mayoraz, E
Aviolat, F
机构
[1] INST DALLE MOLLE INTELLIGENCE ARTIFICIELLE PERCEP, CH-1920 MARTIGNY, SWITZERLAND
[2] SWISS FED INST TECHNOL, DEPT MATH, CH-1015 LAUSANNE, SWITZERLAND
关键词
D O I
10.1142/S0129065796000129
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Quantization of the parameters of a Perceptron is a central problem in hardware implementation of neural networks using a numerical technology A neural model with each weight limited to a small integer range will require little surface of silicon. Moreover, according to Occam's razor principle, better generalisation abilities can be expected from a simpler computational model. The price to pay for these benefits lies in the difficulty to train these kind of networks. This paper proposes essentially two new ideas for constructive training algorithms, and demonstrates their efficiency for the generation of feedforward networks composed of Boolean threshold gates with discrete weights. A proof of the convergence of these algorithms is given. Some numerical experiments have been carried out and the results are presented in terms of the size of the generated networks and of their generalization abilities.
引用
收藏
页码:149 / 166
页数:18
相关论文
共 42 条
  • [1] ALON N, 1991, 8300 RJ IBM RES DIV
  • [2] AMALDI E, 1993, CONSTRUCTIVE METHODS
  • [3] [Anonymous], NEURAL COMPUT
  • [4] Asanovic K., 1991, 2ND P INT C MICR NEU, P9
  • [5] AVIOLAT F, 1993, THESIS ECOLE POLYTEC
  • [6] AVIOLAT F, 1994, P EUR S ART NEUR NET, P123
  • [7] What Size Net Gives Valid Generalization?
    Baum, Eric B.
    Haussler, David
    [J]. NEURAL COMPUTATION, 1989, 1 (01) : 151 - 160
  • [8] Breiman L., 1984, Classification and Regression Trees, DOI DOI 10.2307/2530946
  • [9] BUC FD, 1994, INT J NEURAL SYST, V5, P259
  • [10] DEBENHAM RM, 1988, NEURAL NETWORKS MODE, P752