2021
DOI: 10.3389/fnins.2021.633674
|View full text |Cite
|
Sign up to set email alerts
|

Scaling Equilibrium Propagation to Deep ConvNets by Drastically Reducing Its Gradient Estimator Bias

Abstract: Equilibrium Propagation is a biologically-inspired algorithm that trains convergent recurrent neural networks with a local learning rule. This approach constitutes a major lead to allow learning-capable neuromophic systems and comes with strong theoretical guarantees. Equilibrium propagation operates in two phases, during which the network is let to evolve freely and then “nudged” toward a target; the weights of the network are then updated based solely on the states of the neurons that they connect. The weigh… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
29
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1
1

Relationship

3
6

Authors

Journals

citations
Cited by 30 publications
(29 citation statements)
references
References 15 publications
0
29
0
Order By: Relevance
“…Temporal difference learning is one of the most promising ideas about how backpropagation-like algorithms could be implemented in the brain. It is based on using differences in neuronal activity to approximate top-down error signals 4,[18][19][20][21][22][23][24] . A typical example of such algorithms is contrastive Hebbian learning [25][26][27] , which was proven to be equivalent to backpropagation under certain assumptions 28 .…”
mentioning
confidence: 99%
“…Temporal difference learning is one of the most promising ideas about how backpropagation-like algorithms could be implemented in the brain. It is based on using differences in neuronal activity to approximate top-down error signals 4,[18][19][20][21][22][23][24] . A typical example of such algorithms is contrastive Hebbian learning [25][26][27] , which was proven to be equivalent to backpropagation under certain assumptions 28 .…”
mentioning
confidence: 99%
“…For pooling, we used the max pooling with 2x2 filters and stride 2. The activation function for the convolutional and the fully connected layers was the hard-sigmoid activation function, S(x) = (1+hardtanh(x-1))*0.5, as implemented in (26). The learning rates were: 0.4, 0.028, and 0.025 for the first, second convolutional layer, and for the fully connected output layer, respectively.…”
Section: Convolutional Neural Network (Cifar-10 Dataset)mentioning
confidence: 99%
“…Temporal difference learning is one of the most promising ideas of how backpropagation-like algorithms could be implemented in the brain. It is based on using differences in neuronal activity to approximate top-down error signals (4,(20)(21)(22)(23)(24)(25)(26). A typical example of such algorithms is Contrastive Hebbian Learning (27)(28)(29), which was proven to be equivalent to backpropagation under certain assumptions (30).…”
Section: Introductionmentioning
confidence: 99%
“…Such architectures come with their own technical difficulties. Because they rely on co-located memory and processing resources, standard optimization algorithms can prove impractical, although novel optimization procedures have recently been put forward to overcome this obstacle by meeting their peculiar hardware design [77][78][79][80][81].…”
Section: Introductionmentioning
confidence: 99%