2017
DOI: 10.48550/arxiv.1704.03073
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Data-efficient Deep Reinforcement Learning for Dexterous Manipulation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
66
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
9
1

Relationship

0
10

Authors

Journals

citations
Cited by 49 publications
(67 citation statements)
references
References 0 publications
1
66
0
Order By: Relevance
“…In RL, reward shaping is used to reshape the original reward function to better guide the direction of the gradient update [30]. Prior knowledge about the environment is needed to formalize a reliable reward shaping function to avoid otherwise to bias learning [39].…”
Section: Unbiased Reward Shapingmentioning
confidence: 99%
“…In RL, reward shaping is used to reshape the original reward function to better guide the direction of the gradient update [30]. Prior knowledge about the environment is needed to formalize a reliable reward shaping function to avoid otherwise to bias learning [39].…”
Section: Unbiased Reward Shapingmentioning
confidence: 99%
“…For example, every deadlock that may arise during the previously described optimization scheme should have been predicted, and a corresponding mitigation plan should have been already in place [Palacios-Gasós et al 2016]; otherwise, the robot is going to be stuck in this locally optimal configuration. On top of that, to engineer a multi-term strategy that reflects the task at hand is not always trivial [Popov et al 2017].…”
Section: Related Workmentioning
confidence: 99%
“…Deep Reinforcement learning (D-RL) has been used to learn controllers for a variety of tasks ranging from walking robots [6], [7], [8] to manipulating objects with an arm [9], [10], [11], [12], [13]. Hence reinforcement learning, indeed, offers a way to realize peg-in-hole tasks via random explorations, thereby eliminating the need to hand craft an effective control/policy without using any form of expert data.…”
Section: Related Workmentioning
confidence: 99%