Robotics: Science and Systems VII 2011
DOI: 10.15607/rss.2011.vii.008
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Control a Low-Cost Manipulator using Data-Efficient Reinforcement Learning

Abstract: Abstract-Over the last years, there has been substantial progress in robust manipulation in unstructured environments. The long-term goal of our work is to get away from precise, but very expensive robotic systems and to develop affordable, potentially imprecise, self-adaptive manipulator systems that can interactively perform tasks such as playing with children. In this paper, we demonstrate how a low-cost off-the-shelf robotic system can learn closed-loop policies for a stacking task in only a handful of tri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
182
0

Year Published

2012
2012
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 199 publications
(203 citation statements)
references
References 18 publications
0
182
0
Order By: Relevance
“…Continuous RL has been used for reaching and grasping objects [30][31][32], as well as for the transportation of grasped objects [12,[32][33][34][35]. These methods are driven by feedback from tactile sensing [12,35], Cartesian-and joint-space coordinates [33,34], or both [32]. However, these approaches have not been applied at in-hand manipulation, i.e., changing the object's pose with respect to the hand.…”
Section: Reinforcement Learning For Manipulationmentioning
confidence: 99%
“…Continuous RL has been used for reaching and grasping objects [30][31][32], as well as for the transportation of grasped objects [12,[32][33][34][35]. These methods are driven by feedback from tactile sensing [12,35], Cartesian-and joint-space coordinates [33,34], or both [32]. However, these approaches have not been applied at in-hand manipulation, i.e., changing the object's pose with respect to the hand.…”
Section: Reinforcement Learning For Manipulationmentioning
confidence: 99%
“…However, often many trials are necessary to obtain a good controller. Model-based RL methods potentially use the data more efficient, but it is well known that model bias can strongly degrade the learning performance [4]. Here, a controller might succeed in simulation but fails when applied to the real system, if the model does not describe the complete system dynamics.…”
Section: Approaches In Reinforcement Learningmentioning
confidence: 99%
“…Instead, we include the goal position g as input to the controller, i.e. u = π θ (s, g), as described in [11]. Now, one joint controller is learned for all goal positions.…”
Section: Multiple Start and Goal Statesmentioning
confidence: 99%
See 1 more Smart Citation
“…Reducing overall cost is critical to the commercialization of robotic manipulator technologies [1], [2], especially those envisioned for use in unstructured environments such as households, office spaces and hazardous environments [3], [4]. In order for humanoid robots such as the PR2 [2], NAO [5], and others to reach their target market in households and workplaces, the cost of the robot must decrease.…”
Section: Introductionmentioning
confidence: 99%