2010
DOI: 10.1007/978-3-642-14274-1_8
|View full text |Cite
|
Sign up to set email alerts
|

Reducing the Memory Footprint of Temporal Difference Learning over Finitely Many States by Using Case-Based Generalization

Abstract: Abstract. In this paper we present an approach for reducing the memory footprint requirement of temporal difference methods in which the set of states is finite. We use case-based generalization to group the states visited during the reinforcement learning process. We follow a lazy learning approach; cases are grouped in the order in which they are visited. Any new state visited is assigned to an existing entry in the Q-table provided that a similar state has been visited before. Otherwise a new entry is added… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2012
2012
2012
2012

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
references
References 17 publications
0
0
0
Order By: Relevance