2015
DOI: 10.1007/s10626-015-0216-z
|View full text |Cite
|
Sign up to set email alerts
|

Multiscale Q-learning with linear function approximation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(1 citation statement)
references
References 36 publications
0
1
0
Order By: Relevance
“…Double Q-learning [7] and Speedy Q-learning [4] use two estimates of the Qfunction, for addressing the issues of over-estimation and slow convergence, respectively. A multi-timescale version of the Qlearning algorithm is presented in [8] and its convergence shown using a differential inclusions based analysis. More recently, the Zap Q-learning algorithm was introduced [9], which is a matrix-gain algorithm designed to optimize the asymptotic variance.…”
Section: A Related Workmentioning
confidence: 99%
“…Double Q-learning [7] and Speedy Q-learning [4] use two estimates of the Qfunction, for addressing the issues of over-estimation and slow convergence, respectively. A multi-timescale version of the Qlearning algorithm is presented in [8] and its convergence shown using a differential inclusions based analysis. More recently, the Zap Q-learning algorithm was introduced [9], which is a matrix-gain algorithm designed to optimize the asymptotic variance.…”
Section: A Related Workmentioning
confidence: 99%