Robotics: Science and Systems XIV 2018
DOI: 10.15607/rss.2018.xiv.010
|View full text |Cite
|
Sign up to set email alerts
|

Sim-to-Real: Learning Agile Locomotion For Quadruped Robots

Abstract: Designing agile locomotion for quadruped robots often requires extensive expertise and tedious manual tuning. In this paper, we present a system to automate this process by leveraging deep reinforcement learning techniques. Our system can learn quadruped locomotion from scratch using simple reward signals. In addition, users can provide an open loop reference to guide the learning process when more control over the learned gait is needed. The control policies are learned in a physics simulator and then deploye… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
387
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 520 publications
(406 citation statements)
references
References 34 publications
0
387
0
Order By: Relevance
“…Recent successes in Reinforcement Learning (RL) demonstrate sophisticated walking robot control [1]- [5], yet a large number of policy rollouts need to be collected to reach the required performance level. It is, therefore, common practice to use physics simulators during training and subsequently attempt a sim-to-real transfer [1], [4].…”
Section: Introductionmentioning
confidence: 99%
“…Recent successes in Reinforcement Learning (RL) demonstrate sophisticated walking robot control [1]- [5], yet a large number of policy rollouts need to be collected to reach the required performance level. It is, therefore, common practice to use physics simulators during training and subsequently attempt a sim-to-real transfer [1], [4].…”
Section: Introductionmentioning
confidence: 99%
“…We compare MSO to two baselines: domain randomization (DR) [10] and strategy optimization with projected universal policy (SO-PUP) [11]. We run ARS for 1500 iterations for all methods and we use a two-dimensional latent space for MSO and SO-PUP.…”
Section: A Experiments Setupmentioning
confidence: 99%
“…The first task is to transfer the policy trained in simulation to the real Minitaur robot. Although we use the nonlinear actuator model from Tan et al [10], the reality gap in our case is still large as we use a different version of Minitaur and we do not perform additional system identification.…”
Section: B Adaptation Tasksmentioning
confidence: 99%
See 1 more Smart Citation
“…They can also significantly reduce the time and cost required for repair and maintenance. These features become especially important when testing learning algorithms directly on real hardware [33], [34] where it is essential to have a safe platform to explore various control patterns.…”
Section: Introductionmentioning
confidence: 99%