2021 IEEE International Conference on Robotics and Automation (ICRA) 2021
DOI: 10.1109/icra48506.2021.9561232
|View full text |Cite
|
Sign up to set email alerts
|

LASER: Learning a Latent Action Space for Efficient Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
15
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(15 citation statements)
references
References 18 publications
0
15
0
Order By: Relevance
“…Learning latent action representations is studied in RL [26] and offline RL [27,28]. Allshire et al [26] proposed a conditional variational encoder-decoder (CVAE) model to learn latent action space where generated latent actions are decoded before applying on the original robot state.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Learning latent action representations is studied in RL [26] and offline RL [27,28]. Allshire et al [26] proposed a conditional variational encoder-decoder (CVAE) model to learn latent action space where generated latent actions are decoded before applying on the original robot state.…”
Section: Related Workmentioning
confidence: 99%
“…Learning latent action representations is studied in RL [26] and offline RL [27,28]. Allshire et al [26] proposed a conditional variational encoder-decoder (CVAE) model to learn latent action space where generated latent actions are decoded before applying on the original robot state. van der Pol et al [29] learns latent spaces with equivariant maps to show that optimal policies are equivalent but the application is limited to small discrete systems where value iteration can be applied.…”
Section: Related Workmentioning
confidence: 99%
“…Despite these efforts, today's DRL methods still struggle in longhorizon robotic tasks due to the exploration burden of learning from scratch. A growing amount of work has examined the use of offline data to alleviate the exploration burden in DRL, namely through demonstration-guided RL [17,45,55], learned behavioral priors [54,61] and action spaces [1,2] from demonstrations, and offline RL [14,15,32,39]. While promising, these methods can be difficult to scale up due to the costs of acquiring offline data.…”
Section: Related Workmentioning
confidence: 99%
“…x [2]] and we encourage the agent to specify the reaching parameters x reach to be within a threshold of a set of keypoints P :…”
Section: Facilitating Exploration With Affordancesmentioning
confidence: 99%
See 1 more Smart Citation