Weight Perturbation: An Optimal Architecture and Learning Technique for Analog VLSI Feedforward and Recurrent Multilayer Networks

被引:22
作者
Jabri, Marwan [1 ]
Flower, Barry [1 ]
机构
[1] Univ Sydney, Syst Engn & Design Automat Lab, Sch Elect Engn, Sydney, NSW, Australia
基金
澳大利亚研究理事会;
关键词
D O I
10.1162/neco.1991.3.4.546
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Previous work on analog VLSI implementation of multilayer perceptrons with on-chip learning has mainly targeted the implementation of algorithms like backpropagation. Although backpropagation is efficient, its implementation in analog VLSI requires excessive computational hardware. In this paper we show that, for analog parallel implementations, the use of gradient descent with direct approximation of the gradient using "weight perturbation" instead of backpropagation significantly reduces hardware complexity. Gradient descent by weight perturbation eliminates the need for derivative and bidirectional circuits for on-chip learning, and access to the output states of neurons in hidden layers for off-chip learning. We also show that weight perturbation can be used to implement recurrent networks. A discrete level analog implementation showing the training of an XOR network as an example is described.
引用
收藏
页码:546 / 565
页数:20
相关论文
共 4 条
[1]  
Furman B., 1988, 1ST ANN INNS M BOST, P381
[2]  
Hwang J. N., 1989, Journal of VLSI Signal Processing, V1, P221, DOI 10.1007/BF02427796
[3]   Recurrent Backpropagat ion and the Dynamical Approach to Adaptive Neural Computation [J].
Pineda, Fernando J. .
NEURAL COMPUTATION, 1989, 1 (02) :161-172
[4]   30 YEARS OF ADAPTIVE NEURAL NETWORKS - PERCEPTRON, MADALINE, AND BACKPROPAGATION [J].
WIDROW, B ;
LEHR, MA .
PROCEEDINGS OF THE IEEE, 1990, 78 (09) :1415-1442