2007
DOI: 10.1002/ecjb.20383
|View full text |Cite
|
Sign up to set email alerts
|

An adjustment method of the number of states on Q‐learning segmenting state space adaptively

Abstract: SUMMARYThe results of imposing limitations on the number of states and of promoting the splitting of states in Q-learning are presented. Q-learning is a common reinforcement learning method in which the learning agent autonomously segments the environment states. In situations where the designer of an agent is unable to explicitly provide the agent with the boundaries of states in the environment in which the agent is acting, the agent needs to simultaneously learn while autonomously determining the internal d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
10
0

Year Published

2008
2008
2009
2009

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 7 publications
(7 reference statements)
0
10
0
Order By: Relevance
“…To avoid the curse of dimensionality, there exists modular hierarchical learning [10,11,16] that construct the learning model as the combination of subspaces. Adaptive segmentation [12,13] for constructing the learning model validly corresponding to the environment is also studied. However more effective technique of different approach is also necessary in order to apply reinforcement learning to actual sized problems.…”
Section: Introductionmentioning
confidence: 99%
“…To avoid the curse of dimensionality, there exists modular hierarchical learning [10,11,16] that construct the learning model as the combination of subspaces. Adaptive segmentation [12,13] for constructing the learning model validly corresponding to the environment is also studied. However more effective technique of different approach is also necessary in order to apply reinforcement learning to actual sized problems.…”
Section: Introductionmentioning
confidence: 99%
“…To avoid the curse of dimensionality, there exists modular hierarchical learning [10,11] that construct the learning model as the combination of subspaces. Adaptive segmentation [12,13] for constructing the learning model validly corresponding to the environment is also studied. However more effective technique of different approach is also necessary in order to apply reinforcement learning to actual sized problems.…”
Section: Introductionmentioning
confidence: 99%
“…QLASS (Q-Learning with Adaptive State Segmentation) is one of on-line categorization methods [8,24]. QLASS categorizes percept vectors on the basis of a Voronoi diagram, where each Voronoi cell corresponds to a category.…”
Section: Introductionmentioning
confidence: 99%
“…Since each category is treated as a state in re-inforcement learning, generating too many categories deteriorates the performance of reinforcement learning. In [8], Hamagami and his group proposed some heuristics to reduce the number of categories. In their method, however, the maximum number of categories must be specified although it is difficult to decide the optimal value of it.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation