2022
DOI: 10.1109/lra.2022.3187276
|View full text |Cite
|
Sign up to set email alerts
|

Variable Impedance Skill Learning for Contact-Rich Manipulation

Abstract: Contact-rich manipulation tasks remain a hard problem in robotics that requires interaction with unstructured environments. Reinforcement Learning (RL) is one potential solution to such problems, as it has been successfully demonstrated on complex continuous control tasks. Nevertheless, current state-ofthe-art methods require policy training in simulation to prevent undesired behavior and later domain transfer even for simple skills involving contact. In this paper, we address the problem of learning contact-r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 41 publications
0
6
0
Order By: Relevance
“…The effects of our proposed algorithm are demonstrated through several experiments including improved stability of the algorithm and reduced time consumption of trajectory recurrence in 3C industrial production. There are many other robot precision operations similar to the 3C industry, such as surgical robots [29] and contact‐rich assembly [30]. Our future work will continue to focus on the precise operation of robots in industrial scenarios.…”
Section: Discussionmentioning
confidence: 99%
“…The effects of our proposed algorithm are demonstrated through several experiments including improved stability of the algorithm and reduced time consumption of trajectory recurrence in 3C industrial production. There are many other robot precision operations similar to the 3C industry, such as surgical robots [29] and contact‐rich assembly [30]. Our future work will continue to focus on the precise operation of robots in industrial scenarios.…”
Section: Discussionmentioning
confidence: 99%
“…However, the direction of motion of the robot will not change unless large contact forces are generated, which means that the actual trajectory executed by the robot may not be optimal with respect to energy and time. Furthermore, some scholars [6], [15], [16] learn different impedance parameters from collected expert demonstration trajectories. However, it is difficult to demonstrate trajectories for tasks like ours, i.e., blind maze exploration.…”
Section: B Variable Impedance Control For Contact-rich Tasksmentioning
confidence: 99%
“…Specifically, Martin-Martin et al [7] compared the effect of different action spaces in RL with variable impedance control in end-effector space for contactrich tasks, such as surface-wiping and door-opening. Some work [4], [5], [6] leveraged demonstrations to learn variable impedance parameters as part of the action space. On the other hand, a safe action or policy can be learned with RL to significantly enhance safety during the training stage.…”
Section: Safe Rl For Contact-rich Tasksmentioning
confidence: 99%
See 1 more Smart Citation
“…This model consists of a skill encoder q(z|a), which generates the latent skill actions z ∼ N (µ z , σ z ), and a decoder p dec (a|z), responsible for predicting a sequence of low-level actions a = {a t , • • • , a t+H−1 }, where H ∈ N + denotes the action horizon. Similar to the approach in [28], [29], we train the VAE by optimizing the evidence lower bound (ELBO) given by the equation:…”
Section: A Skill Prior Reinforcement Learningmentioning
confidence: 99%