2010
DOI: 10.1016/j.robot.2010.03.007
|View full text |Cite
|
Sign up to set email alerts
|

Probabilistic Policy Reuse for inter-task transfer learning

Abstract: Policy Reuse is a reinforcement learning technique that efficiently learns a new policy by using past similar learned policies. The Policy Reuse learner improves its exploration by probabilistically including the exploitation of those past policies. Policy Reuse was introduced and previously demonstrated its effectiveness in problems with different reward functions in the same state and action spaces. In this article, we contribute Policy Reuse as transfer learning among different domains. We introduce extende… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
40
0

Year Published

2011
2011
2022
2022

Publication Types

Select...
6
2

Relationship

2
6

Authors

Journals

citations
Cited by 61 publications
(41 citation statements)
references
References 10 publications
0
40
0
Order By: Relevance
“…Policy Reuse has also been succesfully applied in more complex domains, as the Keepaway task in robot soccer, which requires: i) a mapping between tasks that use different state and action spaces; and ii) function approximation methods since the state space is continuous [44,14].…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…Policy Reuse has also been succesfully applied in more complex domains, as the Keepaway task in robot soccer, which requires: i) a mapping between tasks that use different state and action spaces; and ii) function approximation methods since the state space is continuous [44,14].…”
Section: Resultsmentioning
confidence: 99%
“…We assume that we are using a direct RL method to learn the action policy, so we are learning the related Q function. Any RL algorithm can be used to learn the Q function, and Sarsa(λ) and Q(λ) have been applied [13,14].…”
Section: The π-Reuse Exploration Strategymentioning
confidence: 99%
See 1 more Smart Citation
“…When the problem has been learned with m agents, and the next incremental problem with m+1 agents uses a state representation that sensorizes more neighbor agents, we need to use transfer learning techniques as performed for transfer learning in domains like Keepaway [5]. Specifically, a projection is used in order to get a new dataset in the new m + 1-agents problem state space included in R r .…”
Section: Multi-agent It-vqqlmentioning
confidence: 99%
“…This approach is similar in spirit to that of case-based reasoning, where the similarity or reuse function is defined by the parameterized similarity metric. In later work Fernández, García, and Veloso (2010) extend the method so that it can be used in Keepaway as well. Agent and Problem Space.…”
Section: Maze Navigationmentioning
confidence: 99%