2019 International Conference on Robotics and Automation (ICRA) 2019
DOI: 10.1109/icra.2019.8794065
|View full text |Cite
|
Sign up to set email alerts
|

Demonstration-Guided Deep Reinforcement Learning of Control Policies for Dexterous Human-Robot Interaction

Abstract: In this paper, we propose a method for training control policies for human-robot interactions such as handshakes or hand claps via Deep Reinforcement Learning. The policy controls a humanoid Shadow Dexterous Hand, attached to a robot arm. We propose a parameterizable multi-objective reward function that allows learning of a variety of interactions without changing the reward structure. The parameters of the reward function are estimated directly from motion capture data of human-human interactions in order to … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
29
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 35 publications
(29 citation statements)
references
References 21 publications
(70 reference statements)
0
29
0
Order By: Relevance
“…Their framework estimates the speed of the interaction as well, to match the speed of the human. Christen et al [17] use Deep Reinforcement Learning (RL) to learn physical interactions from human-human interactions. They use an imitation reward which helps in learning the intricacies of the interaction.…”
Section: Reaching Phase Of Handshakingmentioning
confidence: 99%
“…Their framework estimates the speed of the interaction as well, to match the speed of the human. Christen et al [17] use Deep Reinforcement Learning (RL) to learn physical interactions from human-human interactions. They use an imitation reward which helps in learning the intricacies of the interaction.…”
Section: Reaching Phase Of Handshakingmentioning
confidence: 99%
“…Our main insight is that sense of touch can guide robots to take actions that result in manipulation of an object. To implement the sense of touch, we extend our state-space with force measurements from tactile sensors positioned at the endeffector [13], [14], [15]. We introduce a touch-based intrinsic reward function to direct exploration towards the states where the robot touches the object.…”
Section: Methodsmentioning
confidence: 99%
“…More similar to our method, [15] and [18] extend their state space with tactile sensing and use it in the reward function. To learn human-robot interactions, [15] propose a specific reward to ensure contact between the human and the robot. [18] uses the feedback of tactile sensors as a penalty to avoid high impacts and hence to learn gentle manipulation.…”
Section: A Tactile Feedbackmentioning
confidence: 99%
See 2 more Smart Citations