Proceedings of the 2020 Genetic and Evolutionary Computation Conference 2020
DOI: 10.1145/3377930.3390214
|View full text |Cite
|
Sign up to set email alerts
|

Evolving inborn knowledge for fast adaptation in dynamic POMDP problems

Abstract: Rapid online adaptation to changing tasks is an important problem in machine learning and, recently, a focus of meta-reinforcement learning. However, reinforcement learning (RL) algorithms struggle in POMDP environments because the state of the system, essential in a RL framework, is not always visible. Additionally, handdesigned meta-RL architectures may not include suitable computational structures for specific learning problems. The evolution of online learning mechanisms, on the contrary, has the ability t… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 14 publications
0
3
0
Order By: Relevance
“…Extending the evolved learning rules to be able to take into account reward signals could greatly improve the model's ability to respond to changes in the environment in an adaptive manner, and it is something we look forward to implementing in future studies. Something similar to this has been explored in other approaches to plastic networks [2,4,13,53], and we expect this to also be a beneficial addition to our approach.…”
Section: Future Directionsmentioning
confidence: 64%
“…Extending the evolved learning rules to be able to take into account reward signals could greatly improve the model's ability to respond to changes in the environment in an adaptive manner, and it is something we look forward to implementing in future studies. Something similar to this has been explored in other approaches to plastic networks [2,4,13,53], and we expect this to also be a beneficial addition to our approach.…”
Section: Future Directionsmentioning
confidence: 64%
“…The continuous control environments are the simple 2D navigation, the half-cheetah direction (Finn et al, 2017) and velocity (Finn et al, 2017) Mujoco (Todorov et al, 2012) based environments and the meta-world ML1 and ML45 environments (Yu et al, 2020). The discrete action environment is a graph navigation environment that supports configurable levels of complexity called the CTgraph (Soltoggio et al, 2021;Ladosz et al, 2021;Ben-Iwhiwhu et al, 2020). The experimental setup focused on investigating the beneficial effect of the proposed neuromodulatory mechanism when augmenting existing meta-RL frameworks (i.e., neuromodulation as complementary tool to meta-RL rather than competing).…”
Section: Results and Analysismentioning
confidence: 99%
“…• Network growth strategies: based on adding parameters to the network in order to allow the learning of new tasks and, at the same time, retain previous knowledge through the old parameters [34], [35], [36]. • Selective plasticity and neuromodulation strategies: methods that allow a subgroup of parameters to be adjusted to the new task while another subgroup is almost or completely frozen [37], [38], [39], [40] . • Generative strategies: inspired by the existence of a dual memory system based on the hippocampus and neocortex in the human brain.…”
Section: Lifelong Machine Learningmentioning
confidence: 99%