2005
DOI: 10.1007/11551188_10
|View full text |Cite
|
Sign up to set email alerts
|

Making Use of Unelaborated Advice to Improve Reinforcement Learning: A Mobile Robotics Approach

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2005
2005
2017
2017

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(17 citation statements)
references
References 7 publications
0
17
0
Order By: Relevance
“…Another solution to this problem is including prior knowledge into the reinforcement learning process. These researches include: the dynamic knowledge-orientated creation of the state space (Hailu, 2001) and the focalization of exploration (Lin, 1992;Millán, Posenato, & Dedieu, 2002;Moreno, Regueiro, & Iglesias, 2004).…”
Section: Related Researchesmentioning
confidence: 99%
See 1 more Smart Citation
“…Another solution to this problem is including prior knowledge into the reinforcement learning process. These researches include: the dynamic knowledge-orientated creation of the state space (Hailu, 2001) and the focalization of exploration (Lin, 1992;Millán, Posenato, & Dedieu, 2002;Moreno, Regueiro, & Iglesias, 2004).…”
Section: Related Researchesmentioning
confidence: 99%
“…The objective of this section is to determine the probability of taking a decision through integrating the above exploitation/exploration policy with the SRL model proposed by Moreno et al (2004). The main idea is as follows (for detail see Moreno et al, 2004).…”
Section: Decision Selection Blockmentioning
confidence: 99%
“…In order to speed up the convergence, Singer and Veloso [14] propose to solve the new problem via inducing the local features of original problem; Hailu and Sommer [15] discuss the effects of different bias information on learning speed by introducing environment information. Moreno et al [16] propose to to introduce prior in supervised reinforcement learning; Lin and Li [17] build a reinforcement learning model based on latent bias; and Fernández and Veloso [18] reuse past learnt bias to supervise the solving of similar tasks. The above approaches use bias to supervise the selection of strategies from actions, can utilize the bias from external environment or past tasks, and thus the learning speed is accelerated.…”
Section: Related Workmentioning
confidence: 99%
“…Additionally, many current advice-taking systems [5], [10] require that the human encode her advice into a scripting or programming language, making it inaccessible to non-technical users.…”
Section: A Advice-taking Agentsmentioning
confidence: 99%