2015
DOI: 10.1109/tnnls.2014.2327636
|View full text |Cite
|
Sign up to set email alerts
|

Self-Organizing Neural Networks Integrating Domain Knowledge and Reinforcement Learning

Abstract: The use of domain knowledge in learning systems is expected to improve learning efficiency and reduce model complexity. However, due to the incompatibility with knowledge structure of the learning systems and real-time exploratory nature of reinforcement learning (RL), domain knowledge cannot be inserted directly. In this paper, we show how self-organizing neural networks designed for online and incremental adaptation can integrate domain knowledge and RL. Specifically, symbol-based domain knowledge is transla… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 42 publications
(12 citation statements)
references
References 23 publications
0
12
0
Order By: Relevance
“…Carroll and Seppi [18] defined similarity in terms of tasks, and proposed several possible task similarity measures based on the transfer time, policy overlap, Q-values, and reward structure respectively to measure the similarity between tasks. Teng [19] proposed to use self-organization neural networks to make the effective use of domain knowledge to reduce model complexity in RL. Through minimizing the reconstruction error of a restricted Boltzmann machine simulating the behavioral dynamics of two compared Markov decision processes, Ammar and Eaton [20] gave a data-driven automated Markov decision process similarity metric.…”
Section: Knowledge Transfer In Rlmentioning
confidence: 99%
See 2 more Smart Citations
“…Carroll and Seppi [18] defined similarity in terms of tasks, and proposed several possible task similarity measures based on the transfer time, policy overlap, Q-values, and reward structure respectively to measure the similarity between tasks. Teng [19] proposed to use self-organization neural networks to make the effective use of domain knowledge to reduce model complexity in RL. Through minimizing the reconstruction error of a restricted Boltzmann machine simulating the behavioral dynamics of two compared Markov decision processes, Ammar and Eaton [20] gave a data-driven automated Markov decision process similarity metric.…”
Section: Knowledge Transfer In Rlmentioning
confidence: 99%
“…Later, in order to solve the problems in the earlier works mentioned above, some automatic similarity estimation works [15][16][17][18][19][20][21] are presented. These works made use data-driven similarity metrics, including the Markov decision process (MDP) similarity metric [20], the Hausdorff metric [21], and the Kantorovich metric [21].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…TD-FALCON has three modes of operation. In the PER-FORM mode, Algorithm 1 is used to select cognitive node J for deriving action choice a for state s. In the LEARN mode, TD-FALCON learns the effect of action choice a on state s. In the INSERT mode, domain knowledge can be assimilated into FALCON [16].…”
Section: A Structure and Operating Modesmentioning
confidence: 99%
“…This means a suitable ρ c3 has to be used. Given that it is difficult to know a priori a suitable ρ c3 , ρ c3 was adapted iteratively [16] using…”
Section: B Bi-directional Adaptationmentioning
confidence: 99%