Robotics: Science and Systems VI 2010
DOI: 10.15607/rss.2010.vi.020
|View full text |Cite
|
Sign up to set email alerts
|

Variable Impedance Control - A Reinforcement Learning Approach

Abstract: Abstract-One of the hallmarks of the performance, versatility, and robustness of biological motor control is the ability to adapt the impedance of the overall biomechanical system to different task requirements and stochastic disturbances. A transfer of this principle to robotics is desirable, for instance to enable robots to work robustly and safely in everyday human environments. It is, however, not trivial to derive variable impedance controllers for practical high DOF robotic tasks. In this contribution, w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
46
0

Year Published

2012
2012
2022
2022

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 57 publications
(48 citation statements)
references
References 20 publications
1
46
0
Order By: Relevance
“…In [4] analytical solutions to optimal control of variable stiffness for maximizing link velocity is reported. Closely related to optimal control is reinforcement learning (RL), which has been used for learning variable impedance policies in [5]. In [6], an EM-based reinforcement learning algorithm initialized by human demonstrations is presented.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In [4] analytical solutions to optimal control of variable stiffness for maximizing link velocity is reported. Closely related to optimal control is reinforcement learning (RL), which has been used for learning variable impedance policies in [5]. In [6], an EM-based reinforcement learning algorithm initialized by human demonstrations is presented.…”
Section: Related Workmentioning
confidence: 99%
“…Lines 8-14 computes the the stiffness matrix based on the current window view of Ξ. Then appropriate damping 5 . Refer to Section III for details on the construction of Φ and Ψ.…”
Section: B Stiffness Adjustment Based On Interactionsmentioning
confidence: 99%
“…We introduce a structure by which P I 2 can learn a discontinuous variable impedance control policy that enables tasks requiring contact, motion, and force control during object interaction. With respect to previous results on variable stiffness control with reinforcement learning [3], here we are not using any policy parameterizations that are based on function approximation. Instead, we represent trajectories and control gains as markov diffusion processes.…”
Section: Figmentioning
confidence: 99%
“…More importantly the optimal control in our approach is an average over sampled controls based on how well they performed on the real system and thus no differentiation of the value functions is performed. With respect to previous work on variable stiffness control with reinforcement learning [4,3], our approach does not use function approximators to represent gains or desired trajectories. This lack of policy parameterization allows us to learn non-smooth trajectories and control gains required for tasks that involve contact, which is particularly important when controlling tendonactuated manipulators.…”
Section: A Optimal Control For Tendon-driven Systemsmentioning
confidence: 99%
See 1 more Smart Citation