2022
DOI: 10.48550/arxiv.2205.07802
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Primacy Bias in Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…As a result, an experience will be used too many times to update the neural networks, which could limit the learning performance to a certain extent [26,27]. In order to relieve the adverse effect of the primacy bias, the weight parameters of the neural networks will be reset regularly [28]. When the outputs of the Q networks become stable, which means the q-function converges, the reset operation will be performed.…”
Section: Training Methodsmentioning
confidence: 99%
“…As a result, an experience will be used too many times to update the neural networks, which could limit the learning performance to a certain extent [26,27]. In order to relieve the adverse effect of the primacy bias, the weight parameters of the neural networks will be reset regularly [28]. When the outputs of the Q networks become stable, which means the q-function converges, the reset operation will be performed.…”
Section: Training Methodsmentioning
confidence: 99%