2024
DOI: 10.1088/2632-2153/ad3ee5
|View full text |Cite
|
Sign up to set email alerts
|

Random feedback alignment algorithms to train neural networks: why do they align?

Dominique Chu,
Florian Bacho

Abstract: Feedback alignment algorithms are an alternative to backpropagation to train neural networks, whereby some of the partial derivatives that are required to compute the gradient are replaced by random terms. This essentially transforms the update rule into a random walk in weight space. Surprisingly, learning still works with those algorithms, including training of deep neural networks. The performance of FA is generally attributed to an alignment of the update of the random walker with the true gradient -… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 17 publications
(22 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?