2013
DOI: 10.1109/tnnls.2013.2270561
|View full text |Cite
|
Sign up to set email alerts
|

Online Selective Kernel-Based Temporal Difference Learning

Abstract: Abstract-In this paper, an online selective kernel-based temporal difference (OSKTD) learning algorithm is proposed to deal with large scale and/or continuous reinforcement learning problems. OSKTD includes two online procedures: online sparsification and parameter updating for the selective kernelbased value function. A new sparsification method (i.e., a kernel distance-based online sparsification method) is proposed based on selective ensemble learning, which is computationally less complex compared with oth… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2015
2015
2019
2019

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 24 publications
(7 citation statements)
references
References 44 publications
0
7
0
Order By: Relevance
“…After 43 trials, we count the number of trials which received a positive reward, and the success rate is averaged over 50 Monte Carlo runs. The performance of the Q -KTD algorithm is compared with Q -learning via time delayed neural net ( Q -TDNN) and the online selective kernel-based temporal difference learning algorithm ( Q -OSKTD) [23] in Figure 9. Note that TDNN is a conventional approach to function approximation and has already been applied to RLBMI experiments for neural decoding [1, 2].…”
Section: Experimental Results On Neural Decodingmentioning
confidence: 99%
See 4 more Smart Citations
“…After 43 trials, we count the number of trials which received a positive reward, and the success rate is averaged over 50 Monte Carlo runs. The performance of the Q -KTD algorithm is compared with Q -learning via time delayed neural net ( Q -TDNN) and the online selective kernel-based temporal difference learning algorithm ( Q -OSKTD) [23] in Figure 9. Note that TDNN is a conventional approach to function approximation and has already been applied to RLBMI experiments for neural decoding [1, 2].…”
Section: Experimental Results On Neural Decodingmentioning
confidence: 99%
“…These methods, known as kernel sparsification methods, can be applied to the KTD algorithm to control the growth of the terms in the function expansion, also known as filter size. Popular examples of kernel sparsification methods are the approximate linear dependence (ALD) [19], Surprise criterion [32], Quantization approach [21], and the kernel distance based method [23]. The main idea of sparsification is to only consider a reduced set of samples, called the dictionary, to represent the function of interest.…”
Section: Online Sparsificationmentioning
confidence: 99%
See 3 more Smart Citations