2022
DOI: 10.3389/fncom.2022.980613
|View full text |Cite
|
Sign up to set email alerts
|

Combining backpropagation with Equilibrium Propagation to improve an Actor-Critic reinforcement learning framework

Abstract: Backpropagation (BP) has been used to train neural networks for many years, allowing them to solve a wide variety of tasks like image classification, speech recognition, and reinforcement learning tasks. But the biological plausibility of BP as a mechanism of neural learning has been questioned. Equilibrium Propagation (EP) has been proposed as a more biologically plausible alternative and achieves comparable accuracy on the CIFAR-10 image classification task. This study proposes the first EP-based reinforceme… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 29 publications
(39 reference statements)
0
2
0
Order By: Relevance
“…MLP continuously learns to produce increasingly accurate predictions by repeatedly modifying the weights depending on the computed gradients. 99 BP-MLP has also achieved 81% accuracy for their prediction model for four tasters' parameters (leaf quality, infusion, liquor, and aroma) used for their black tea samples to correlate with sensory analysis responses, which was lower than the 90% obtained with the PNN. 98 According to the authors, greater PNN values might be linked to their dataset size and also the idea that the PNN is derived from a probability base that uses a previous sample to move forward the model output, which depending on the variations in your sample size favors more accurate values.…”
Section: Data Treatment Used For Lc-e-nose Based On Machine Learningmentioning
confidence: 87%
“…MLP continuously learns to produce increasingly accurate predictions by repeatedly modifying the weights depending on the computed gradients. 99 BP-MLP has also achieved 81% accuracy for their prediction model for four tasters' parameters (leaf quality, infusion, liquor, and aroma) used for their black tea samples to correlate with sensory analysis responses, which was lower than the 90% obtained with the PNN. 98 According to the authors, greater PNN values might be linked to their dataset size and also the idea that the PNN is derived from a probability base that uses a previous sample to move forward the model output, which depending on the variations in your sample size favors more accurate values.…”
Section: Data Treatment Used For Lc-e-nose Based On Machine Learningmentioning
confidence: 87%
“…In future work, we plan to apply models with adaptation to neuronal data analyses [ 43–47 ] and to reinforcement learning tasks [ 48 , 49 ]. Regularization effect of adaptation may help to improve training networks with BP.…”
Section: Discussionmentioning
confidence: 99%
“…These agents actively engage with their surroundings, with deep neural networks guiding their decisions and actions ( Osband et al, 2016 ). Among the various algorithms in DRL, the Actor-Critic (AC) algorithm is recognized as a prominent and effective approach ( Kubo, Chalmers & Luczak, 2022 ).…”
Section: Introductionmentioning
confidence: 99%