1992
DOI: 10.1109/72.125861
|View full text |Cite
|
Sign up to set email alerts
|

Using random weights to train multilayer networks of hard-limiting units

Abstract: A gradient descent algorithm suitable for training multilayer feedforward networks of processing units with hard-limiting output functions is presented. The conventional backpropagation algorithm cannot be applied in this case because the required derivatives are not available. However, if the network weights are random variables with smooth distribution functions, the probability of a hard-limiting unit taking one of its two possible values is a continuously differentiable function. In the paper, this is used… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0
3

Year Published

1993
1993
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 44 publications
(22 citation statements)
references
References 10 publications
0
18
0
3
Order By: Relevance
“…Threshold activation functions can be realized with analog comparators. Very few hardlimiting function, multilayer network training algorithms have been developed [15,16,20,21,22]. A multilayer extension of the LMS, known as MR2, was chosen for ease of coding [1 6].…”
Section: Algorithm Selectionmentioning
confidence: 99%
“…Threshold activation functions can be realized with analog comparators. Very few hardlimiting function, multilayer network training algorithms have been developed [15,16,20,21,22]. A multilayer extension of the LMS, known as MR2, was chosen for ease of coding [1 6].…”
Section: Algorithm Selectionmentioning
confidence: 99%
“…the means and standard deviations of the appropriate Gaussian distributions. 33,34 In the reversible generalisation, where each neuron is replaced by a permutation matrix, we find that the output is no longer a function of the inputs and continuous weights, but rather of the inputs and a discrete set of permutation matrices. However, in the generalisation to unitaries, for a gate with n inputs and outputs, there exist an infinite number of unitaries, in contrast with the discrete set of permutation matrices.…”
mentioning
confidence: 97%
“…Bartlett in [1] introduced another approach by defining the weights as random variables with smooth distribution functions and proposed an algorithm that uses an approach that is similar to BP to adjust the parameters of the weights' distributions. In [4], Corwin suggested to train NDAs with progressively steeper analog functions to facilitate training.…”
Section: Training Methods For Network With Discrete Activationsmentioning
confidence: 99%
“…Various modifications of the gradient descent have been presented to train NDAs [1,3,4,6,14,17]. However, these methods require to a certain degree, depending on the method, that the learning task should be static.…”
Section: Introductionmentioning
confidence: 99%