2018 IEEE International Conference on Agents (ICA) 2018
DOI: 10.1109/agents.2018.8460067
|View full text |Cite
|
Sign up to set email alerts
|

Probabilistic Guided Exploration for Reinforcement Learning in Self-Organizing Neural Networks

Abstract: Exploration is essential in reinforcement learning, which expands the search space of potential solutions to a given problem for performance evaluations. Specifically, carefully designed exploration strategy may help the agent learn faster by taking the advantage of what it has learned previously. However, many reinforcement learning mechanisms still adopt simple exploration strategies, which select actions in a pure random manner among all the feasible actions. In this paper, we propose novel mechanisms to im… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 5 publications
(7 reference statements)
0
2
0
Order By: Relevance
“…Another approach of this type could be reducing states for exploration based on some predefined metric. An approach using the adaptive resonance theorem (ART) [150] was presented in [151] and was later extended in [152]. In ART, knowledge about actions can be split into: (i) positive chunk which leads to positive rewards, (ii) negative chunk which leads to negative results, and (iii) empty chunk which is not yet taken.…”
Section: Exploration Parameters Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Another approach of this type could be reducing states for exploration based on some predefined metric. An approach using the adaptive resonance theorem (ART) [150] was presented in [151] and was later extended in [152]. In ART, knowledge about actions can be split into: (i) positive chunk which leads to positive rewards, (ii) negative chunk which leads to negative results, and (iii) empty chunk which is not yet taken.…”
Section: Exploration Parameters Methodsmentioning
confidence: 99%
“…In this approach, the action is randomly chosen from positive and no chunks; thus, the agent is exploring either new things or ones with the positive reward. Wang et al [152] extended this to include the probability of selecting the remaining actions based on how well they are known.…”
Section: Exploration Parameters Methodsmentioning
confidence: 99%