2021
DOI: 10.48550/arxiv.2109.13362
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

FastMimic: Model-based Motion Imitation for Agile, Diverse and Generalizable Quadrupedal Locomotion

Abstract: Robots operating in human environments need a variety of skills, like slow and fast walking, turning, and sidestepping. However, building robot controllers that can exhibit such a large range of behaviors is challenging, and unsolved. We present an approach that uses a model-based controller for imitating different animal gaits without requiring any realworld fine-tuning. Unlike previous works that learn one policy per motion, we present a unified controller which is capable of generating four different animal… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
4

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 37 publications
(66 reference statements)
0
4
0
Order By: Relevance
“…While NLP-solver-based MPC often only applies u * to the system, resulting in open-loop MPC, the DDP-based MPC addtionally applies K * to the system, resulting in feedback MPC. The feedback MPC can account for more policy lags and earn more robustness [9], [26]. Further, during replanning, most single-shooting NLP solvers use u * from the previous planning to warm start the current optimization.…”
Section: B Schemes Of Maintaining Fast Convergencementioning
confidence: 99%
“…While NLP-solver-based MPC often only applies u * to the system, resulting in open-loop MPC, the DDP-based MPC addtionally applies K * to the system, resulting in feedback MPC. The feedback MPC can account for more policy lags and earn more robustness [9], [26]. Further, during replanning, most single-shooting NLP solvers use u * from the previous planning to warm start the current optimization.…”
Section: B Schemes Of Maintaining Fast Convergencementioning
confidence: 99%
“…Thus, we also experiment with another model-predictive control (MPC) dynamic controller from [38], which commands joint torques directly. This controller has been applied to real-world A1 robot [52,53] and shows better tracking of desired velocities for our test robots, as compared to the Raibert controller from [22]. However, MPC is prohibitively slow and cannot be used for training RL policies.…”
Section: Kinematic and Dynamic Control For Visual Navigationmentioning
confidence: 99%
“…However, while controllers behave well in idealized simulated environments, they often struggle when transferred to the real world, exhibiting infeasible motor-control behaviors due to the difference between simulation and real-world, which is often referred to as the reality gap. Some approaches propose to address the reality gap with conventional optimization methods such as MPC, allowing the policy to adjust on the real-robot [30,68,40,76]. On the other hand, others have investigated methods that leverage real-world data, such as learning on real robots [22,21,62], identifying system parameters [29], or adapting policy behaviors [53,83,36].…”
Section: Related Work a Legged Robot Controlmentioning
confidence: 99%