2014 IEEE/RSJ International Conference on Intelligent Robots and Systems 2014
DOI: 10.1109/iros.2014.6942745
|View full text |Cite
|
Sign up to set email alerts
|

Latent space policy search for robotics

Abstract: Abstract-Learning motor skills for robots is a hard task. In particular, a high number of degrees-of-freedom in the robot can pose serious challenges to existing reinforcement learning methods, since it leads to a highdimensional search space. However, complex robots are often intrinsically redundant systems and, therefore, can be controlled using a latent manifold of much smaller dimensionality. In this paper, we present a novel policy search method that performs efficient reinforcement learning by uncovering… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
22
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 23 publications
(23 citation statements)
references
References 24 publications
0
22
0
Order By: Relevance
“…While most approaches found in literature perform DR in the joint space [27], [25], [26], for comparison purposes we also derived DR in the parameter space. To do so, the procedure is equivalent to that of the previous subsections, with the exception that now the parameter r disappears and we introduce the parameter M f ≤ dN f , indicating the total number of Gaussian parameters used.…”
Section: Dimensionality Reduction In the Parameter Space (Pdr-dmp)mentioning
confidence: 99%
See 2 more Smart Citations
“…While most approaches found in literature perform DR in the joint space [27], [25], [26], for comparison purposes we also derived DR in the parameter space. To do so, the procedure is equivalent to that of the previous subsections, with the exception that now the parameter r disappears and we introduce the parameter M f ≤ dN f , indicating the total number of Gaussian parameters used.…”
Section: Dimensionality Reduction In the Parameter Space (Pdr-dmp)mentioning
confidence: 99%
“…Different variants of the proposed latent space DMP representation have been tested, as well as an EM-based approach [27] adapted to the DMPs. We used episodic REPS in all the experiments and, therefore, timestep learning methods like [25] were not included in the experimentation. The application of the proposed methods does not depend on the REPS algorithm, as they can be implemented with any PS procedure using Gaussian weighted maximum likelihood estimation for reevaluating the policy parameters, such as for example PI2 [19], [20], [21].…”
Section: Experimentationmentioning
confidence: 99%
See 1 more Smart Citation
“…In [1], the authors proposed to reduce the dimensionality of a ProMP by performing linear DR in the space of degrees of freedom (DoF) of the robot, to reduce the number of DoFs from d to r. This had the impact of reducing the dimensionality of the parameter vector ω from dN f to rN f , with r < d, N f being the number of Gaussian kernels used per DoF. While reducing the dimensionality in the robots' DoF has advantages such as a better qualitative understanding of the task, a smaller linear projection matrix which is easier to estimate, and it is also used in other approaches [18]; However, here we propose to reduce the dimensionality in the space of the Gaussian weight vectors ω. This variation is introduced to then build a GMM in the common latent parameter space and has the advantage of fine-tuning the dimensionality of the latent space, given that we can encode actions that are more different, without loosing too much information.…”
Section: A Dimensionality Reduction Of Prompsmentioning
confidence: 99%
“…Dimensionality reduction over the DoF of robots is a common approach for grasping and hand motion [6], [7]. However, it has been less used for arm robot skills.…”
Section: Introductionmentioning
confidence: 99%