2022
DOI: 10.1613/jair.1.13554
|View full text |Cite
|
Sign up to set email alerts
|

Autotelic Agents with Intrinsically Motivated Goal-Conditioned Reinforcement Learning: A Short Survey

Abstract: Building autonomous machines that can explore open-ended environments, discover possible interactions and build repertoires of skills is a general objective of artificial intelligence. Developmental approaches argue that this can only be achieved by autotelic agents: intrinsically motivated learning agents that can learn to represent, generate, select and solve their own problems. In recent years, the convergence of developmental approaches with deep reinforcement learning (RL) methods has been leading to the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
61
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 55 publications
(87 citation statements)
references
References 68 publications
(93 reference statements)
0
61
0
Order By: Relevance
“…control instructions and act toward various goals (Veeriah et al, 2018). Based on universal value function approximators (UVFAs) (Schaul et al, 2015), goal-conditioned RL (GCRL) (Colas et al, 2022) is proposed to accomplish these tasks by leveraging the goalconditioned value network and policy network. The RL agent is optimized by goal-labeled trajectories with goal-specific rewards.…”
Section: Figurementioning
confidence: 99%
“…control instructions and act toward various goals (Veeriah et al, 2018). Based on universal value function approximators (UVFAs) (Schaul et al, 2015), goal-conditioned RL (GCRL) (Colas et al, 2022) is proposed to accomplish these tasks by leveraging the goalconditioned value network and policy network. The RL agent is optimized by goal-labeled trajectories with goal-specific rewards.…”
Section: Figurementioning
confidence: 99%
“…In reinforcement learning (RL), the reward function specifies the external, environmentally-defined signal the agent attempts to maximize: formally, it maps each state and each action to the scalar reward the agent experiences for taking an action at a particular state. Agent-generated goals are used to supplement the environment's reward function and provide an alternative signal that can guide exploration (see Colas et al, 2020b; we will omit discussion of other exploration approaches, see Weng, 2020 for a review). Most current approaches generate goals that can be evaluated on only a single world state, and as such, fall far short of the richness of real-world goals (Colas et al, 2020b, section 7.1).…”
Section: Related Workmentioning
confidence: 99%
“…In the larger context, acquisition is an important part of simulating the development of intelligent behavior as well. So-called autotelic agents serve as a meta-approach to acquisition by modeling open-ended adaptive behaviors [19]. Developmental connectionist models enable the developmental trajectories and transitions, critical periods, and the process of learning [20,21].…”
Section: Simulation Of Interconnected Developmental Systemsmentioning
confidence: 99%