2013
DOI: 10.1016/j.neucom.2012.08.039
|View full text |Cite
|
Sign up to set email alerts
|

Transferring task models in Reinforcement Learning agents

Abstract: The main objective of Transfer Learning is to reuse knowledge acquired in a previous learned task, in order to enhance the learning procedure in a new and more complex task. Transfer learning comprises a suitable solution for speeding up the learning procedure in Reinforcement Learning tasks. This work proposes a novel method for transferring models to Reinforcement Learning agents. The models of the transition and reward functions of a source task, will be transferred to a relevant but different target task. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
14
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 18 publications
(14 citation statements)
references
References 21 publications
(34 reference statements)
0
14
0
Order By: Relevance
“…The experiments verified the initial claim that L3 outperforms its non transfer-learning versions, and also state-of-the-art transfer learning algorithms. The results show that L3 outperforms traditional RL algorithms, such as Q-Learning and SARSA(λ), HARL algorithms such as the HAQL and HA-SARSA(λ), and TL algorithms such as Taylor's MASTER (Taylor, 2008) and TiMRLA Value-Addition algorithms (Fachantidis et al, 2013). It is worth pointing out that, in contrast to L3, HAQL and HA-SARSA(λ) presuppose user-defined domain knowledge.…”
Section: Discussionmentioning
confidence: 95%
See 3 more Smart Citations
“…The experiments verified the initial claim that L3 outperforms its non transfer-learning versions, and also state-of-the-art transfer learning algorithms. The results show that L3 outperforms traditional RL algorithms, such as Q-Learning and SARSA(λ), HARL algorithms such as the HAQL and HA-SARSA(λ), and TL algorithms such as Taylor's MASTER (Taylor, 2008) and TiMRLA Value-Addition algorithms (Fachantidis et al, 2013). It is worth pointing out that, in contrast to L3, HAQL and HA-SARSA(λ) presuppose user-defined domain knowledge.…”
Section: Discussionmentioning
confidence: 95%
“…L3 was compared with other state-of-the-art transfer learning algorithms: L3 was compared with the the TiMRLA Value-Addition algorithm transfer learning algorithm (Fachantidis et al, 2013) and Taylor's MASTER algorithm (Taylor, 2008) on the Mountain Car experiment (Section 5.1). We did not find any competing algorithm in the literature to compare the Humanoid Robot Stabilisation experiment (presented in Section 5.2), which was a domain included in this paper to illustrate the generality of the algorithm proposed.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Fachantidis et al [10] transfer the model of the transition and reward functions of a source task to a relevant but different target task, and then the agent takes a hybrid approach, implementing both model-free and model-based learning. But all the two methods are limited by the need of inter-task mapping.…”
Section: Related Workmentioning
confidence: 99%