1991
DOI: 10.1162/neco.1991.3.4.546
|View full text |Cite
|
Sign up to set email alerts
|

Weight Perturbation: An Optimal Architecture and Learning Technique for Analog VLSI Feedforward and Recurrent Multilayer Networks

Abstract: Previous work on analog VLSI implementation of multilayer perceptrons with on-chip learning has mainly targeted the implementation of algorithms like backpropagation. Although backpropagation is efficient, its implementation in analog VLSI requires excessive computational hardware. In this paper we show that, for analog parallel implementations, the use of gradient descent with direct approximation of the gradient using "weight perturbation" instead of backpropagation significantly reduces hardware complexity.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
52
0

Year Published

1994
1994
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 58 publications
(52 citation statements)
references
References 3 publications
0
52
0
Order By: Relevance
“…In weight perturbation (Jabri and Flower, 1991;Alspector et al, 1993;Flower and Jabri, 1993;Kirk et al, 1993;Cauwenberghs, 1993) the gradient ' is approximated using only the globally broadcast result of the computation of E(© ). This is done by adding a random zero-mean perturbation vector ∆© to © repeatedly and approxmating the resulting change in error by…”
Section: Weight Perturbationmentioning
confidence: 99%
“…In weight perturbation (Jabri and Flower, 1991;Alspector et al, 1993;Flower and Jabri, 1993;Kirk et al, 1993;Cauwenberghs, 1993) the gradient ' is approximated using only the globally broadcast result of the computation of E(© ). This is done by adding a random zero-mean perturbation vector ∆© to © repeatedly and approxmating the resulting change in error by…”
Section: Weight Perturbationmentioning
confidence: 99%
“…Spike timing dependent plasticity [77] is a recent learning scheme that exploits the precise timing between pre and post synaptic firing events in the learning rule. Numerous adaptations of these major learning rules have been proposed, as well as hardware-friendly rules, such as weight perturbation [78].…”
Section: Learningmentioning
confidence: 99%
“…The stochastic gradient descent algorithm, also called the weight perturbation algorithm (Jabri and Flower, 1992), is a simple method for descending the gradient of a noisy objective function. The algorithm proceeds as follows.…”
Section: Risk-sensitive Stochastic Gradient Descentmentioning
confidence: 99%