2013
DOI: 10.3389/fnbot.2013.00022
|View full text |Cite
|
Sign up to set email alerts
|

Which is the best intrinsic motivation signal for learning multiple skills?

Abstract: Humans and other biological agents are able to autonomously learn and cache different skills in the absence of any biological pressure or any assigned task. In this respect, Intrinsic Motivations (i.e., motivations not connected to reward-related stimuli) play a cardinal role in animal learning, and can be considered as a fundamental tool for developing more autonomous and more adaptive artificial agents. In this work, we provide an exhaustive analysis of a scarcely investigated problem: which kind of IM reinf… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
49
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
4
3
2
1

Relationship

3
7

Authors

Journals

citations
Cited by 53 publications
(55 citation statements)
references
References 34 publications
1
49
0
Order By: Relevance
“…The learning progress in achieving a goal can be used as a transient reward so that the system focuses on tasks where it is learning the most, moving to other ones when the task-related skill has been completely learnt or when more promising activities come at hand [11], [12]. This strategy allows the learning of multiple separated skills, and possibly a dynamical transfer of knowledge between tasks that require similar policies [13].…”
Section: Introductionmentioning
confidence: 99%
“…The learning progress in achieving a goal can be used as a transient reward so that the system focuses on tasks where it is learning the most, moving to other ones when the task-related skill has been completely learnt or when more promising activities come at hand [11], [12]. This strategy allows the learning of multiple separated skills, and possibly a dynamical transfer of knowledge between tasks that require similar policies [13].…”
Section: Introductionmentioning
confidence: 99%
“…The matching value is used to determine the probability distribution over the goals that supports the selection of the goal to pursue (number 2 in the figure). The probability distribution is generated on the basis of intrinsic motivations, for example related to competence (e.g., the goals with a lower competence, or with a higher competence-improvement, have a higher probability of selection; Santucci et al, 2013).…”
Section: Agentmentioning
confidence: 99%
“…Reinforcement Learning Barto et al, 2004;Simşek and Barto, 2006;Schembri et al, 2007;Sequeira et al, 2011;Kompella et al, 2012;Baldassarre and Mirolli, 2013;Metzen and Kirchner, 2013;Di Nocera et al, 2014;Frank et al, 2015;Hester and Stone, 2015 Deep Learning Mohamed and Rezende, 2015;Kulkarni et al, 2016;Achiam and Sastry, 2017;Zhelo et al, 2018 Hierarchical Structure Schembri et al, 2007;Baranes and Oudeyer, 2010;Baldassarre and Mirolli, 2013;Santucci et al, 2013;Frank et al, 2015;Kulkarni et al, 2016 Active Learning Baranes and Oudeyer, 2009, 2010Kompella et al, 2017;Pathak et al, 2017 Motion Planning Frank et al, 2015 Affordance Discovery Hart et al, 2008;Hart, 2009 Goal Discovery/Goal Generation In Reinforcement Learning (RL), an agent learns from experience as it deals with a sequential decision problem. The agent interacts with an "environment" which contains a "critic" that provides the agent with rewards by evaluating the behavior.…”
Section: Settings Referencesmentioning
confidence: 99%