2020
DOI: 10.48550/arxiv.2011.12428
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Align, then memorise: the dynamics of learning with feedback alignment

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
3
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…While RBP addresses the Symmetry problem, by itself it does not address the other problems. Furthermore, while RBP simulations have succeeded on the MNIST and CIFAR benchmark data sets, it has been noted that RBP algorithms do not work well with convolutional layers [Bartunov et al, 2018, Moskovitz et al, 2018, Refinetti et al, 2020. A few methods have been proposed to address this apparent weakness of RBP algorithms, however, most of them are not biologically plausible.…”
Section: Random Backpropagationmentioning
confidence: 99%
See 1 more Smart Citation
“…While RBP addresses the Symmetry problem, by itself it does not address the other problems. Furthermore, while RBP simulations have succeeded on the MNIST and CIFAR benchmark data sets, it has been noted that RBP algorithms do not work well with convolutional layers [Bartunov et al, 2018, Moskovitz et al, 2018, Refinetti et al, 2020. A few methods have been proposed to address this apparent weakness of RBP algorithms, however, most of them are not biologically plausible.…”
Section: Random Backpropagationmentioning
confidence: 99%
“…After building the twin Tourbillon architectures and training their circular autoencoders by recirculation, we apply fine-tuning as described in Section 5.1 to perform random backpropagation using the channel provided by the decoders' connections. Empirically, we found that it is necessary to have different learning rate schedules for different layers during the fine-tuning phase to ensure a form of asynchronous learning among layers(see also Refinetti et al [2020]). A summary of the results comparing the various approaches is given in Table 1 (first row), and shows that the performance of the Tourbillon twin approach is comparable to backpropagation.…”
Section: Tourbillonizationmentioning
confidence: 99%
“…Learning with DFA is enabled by an alignment process, wherein the forward weights learn a configuration enabling DFA to approximate BP updates [42].…”
Section: Our Defensementioning
confidence: 99%
“…For ease of notation, we introduce it for fully connected networks but it generalizes to convolutional networks, transformers and other architectures [16]. It has been theoretically studied in [20,26]. Note that in the following, we incorporate the bias terms in the weight matrices.…”
Section: Learning With Direct Feedback Alignment (Dfa)mentioning
confidence: 99%