2021
DOI: 10.48550/arxiv.2110.02639
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On The Transferability of Deep-Q Networks

Abstract: Transfer Learning (TL) is an efficient machine learning paradigm that allows overcoming some of the hurdles that characterize the successful training of deep neural networks, ranging from long training times to the needs of large datasets. While exploiting TL is a well established and successful training practice in Supervised Learning (SL), its applicability in Deep Reinforcement Learning (DRL) is rarer.In this paper, we study the level of transferability of three different variants of Deep-Q Networks on popu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 26 publications
(29 reference statements)
0
2
0
Order By: Relevance
“…Actor-critic methods [69,94] are hybrid approaches that use policy-based methods to improve a policy while also evaluating it by estimating its corresponding value function. Several studies, including [22,95], investigated the adaptability of value-based algorithms to environmental changes. They trained a value function on a source task and then transferred the value function's parameters to a new task that differed from the source task in transition dynamics.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Actor-critic methods [69,94] are hybrid approaches that use policy-based methods to improve a policy while also evaluating it by estimating its corresponding value function. Several studies, including [22,95], investigated the adaptability of value-based algorithms to environmental changes. They trained a value function on a source task and then transferred the value function's parameters to a new task that differed from the source task in transition dynamics.…”
Section: Related Workmentioning
confidence: 99%
“…MF methods enable quick and computationally efficient action selection at decision time. However, as recent research [22,23,24,25] has shown, the adaptability of MF frameworks to environmental changes does not appear to be promising. This is due to the fact that an MF agent cannot adapt cached values of all states to changes in the environment.…”
Section: Introductionmentioning
confidence: 99%