2021
DOI: 10.48550/arxiv.2106.07278
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Which Mutual-Information Representation Learning Objectives are Sufficient for Control?

Kate Rakelly,
Abhishek Gupta,
Carlos Florensa
et al.

Abstract: Mutual information maximization provides an appealing formalism for learning representations of data. In the context of reinforcement learning (RL), such representations can accelerate learning by discarding irrelevant and redundant information, while retaining the information necessary for control. Much of the prior work on these methods has addressed the practical difficulties of estimating mutual information from samples of high-dimensional observations, while comparatively less is understood about which mu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

1
0
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 39 publications
(67 reference statements)
1
0
0
Order By: Relevance
“…Since this result holds for all Q-functions, it also holds for the optimal Q * and Q * . This result extends sufficiency results in prior works Rakelly et al (2021) to a setting with stochastic encoders and KL-divergence losses on forward dynamics and reward, as opposed to Wasserstein metrics (Gelada et al, 2019), and bisimulation metrics (Zhang et al, 2020).…”
Section: A4 Value Difference Resultssupporting
confidence: 81%
“…Since this result holds for all Q-functions, it also holds for the optimal Q * and Q * . This result extends sufficiency results in prior works Rakelly et al (2021) to a setting with stochastic encoders and KL-divergence losses on forward dynamics and reward, as opposed to Wasserstein metrics (Gelada et al, 2019), and bisimulation metrics (Zhang et al, 2020).…”
Section: A4 Value Difference Resultssupporting
confidence: 81%