2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2015
DOI: 10.1109/iros.2015.7354296
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement learning vs human programming in tetherball robot games

Abstract: Reinforcement learning of motor skills is an important challenge in order to endow robots with the ability to learn a wide range of skills and solve complex tasks. However, comparing reinforcement learning against human programming is not straightforward. In this paper, we create a motor learning framework consisting of state-of-the-art components in motor skill learning and compare it to a manually designed program on the task of robot tetherball. We use dynamical motor primitives for representing the robot's… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
6
1
1

Relationship

2
6

Authors

Journals

citations
Cited by 13 publications
(18 citation statements)
references
References 10 publications
0
18
0
Order By: Relevance
“…To this end, seminal work in robotics has proposed using dynamical systems to model actions and trajectories, in a continuous space. Dynamic Movement Primitives (DMPs) [12,33,29] have been widely used to perform diverse, dynamic tasks such as table tennis [22], panckake flipping [16] or tether-ball [25]. They are able to model smooth, natural motions, and have in fact been used to inspire many policy learning schemes [8,5,4,40,11,7].…”
Section: Related Workmentioning
confidence: 99%
“…To this end, seminal work in robotics has proposed using dynamical systems to model actions and trajectories, in a continuous space. Dynamic Movement Primitives (DMPs) [12,33,29] have been widely used to perform diverse, dynamic tasks such as table tennis [22], panckake flipping [16] or tether-ball [25]. They are able to model smooth, natural motions, and have in fact been used to inspire many policy learning schemes [8,5,4,40,11,7].…”
Section: Related Workmentioning
confidence: 99%
“…To this end, seminal work in robotics has proposed using dynamical systems to model actions and trajectories, in a continuous space. Dynamic Movement Primitives (DMPs) [12,34,30] have been widely used to perform diverse, dynamic tasks such as table tennis [23], panckake flipping [16] or tether-ball [26]. They are able to model smooth, natural motions, and have in fact been used to inspire many policy learning schemes [8,5,4,41,11,7].…”
Section: Related Workmentioning
confidence: 99%
“…Approaches maximizing such functions are commonly referred to as black-box optimizers. Such a formulation of the RL problem has been shown to be effective for learning or adjusting behavioral policies in robotic scenarios (Kupcsik et al, 2013;Parisi et al, 2015), especially when carefully designing the policy π(a|s, θ) to ensure safe behavior while using only a lowdimensional parameterization θ. One reason for the effectiveness is that the exploration of the algorithm is performed on the parameters θ instead of on the actions a.…”
Section: Application To Episodic Reinforcement Learningmentioning
confidence: 99%
“…If the policy parameterization is well-chosen for the task, this form of exploration can be much more effective. The contextual relative entropy policy search (C-REPS) algorithm (Neumann, 2011;Kupcsik et al, 2013;Parisi et al, 2015) frames the maximization of ( 15) over a task distribution µ(c) as a repeated entropy-regularized optimization max q(θ,c)…”
Section: Application To Episodic Reinforcement Learningmentioning
confidence: 99%