2022
DOI: 10.48550/arxiv.2206.09674
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

EAGER: Asking and Answering Questions for Automatic Reward Shaping in Language-guided RL

Abstract: Reinforcement learning (RL) in long horizon and sparse reward tasks is notoriously difficult and requires a lot of training steps. A standard solution to speed up the process is to leverage additional reward signals, shaping it to better guide the learning process. In the context of language-conditioned RL, the abstraction and generalisation properties of the language input provide opportunities for more efficient ways of shaping the reward. In this paper, we leverage this idea and propose an automated reward … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 21 publications
(27 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?