2022
DOI: 10.48550/arxiv.2208.07363
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control

Abstract: Simulated humanoids are an appealing research domain due to their physical capabilities. Nonetheless, they are also challenging to control, as a policy must drive an unstable, discontinuous, and high-dimensional physical system. One widely studied approach is to utilize motion capture (MoCap) data to teach the humanoid agent low-level skills (e.g., standing, walking, and running) that can then be re-used to synthesize high-level behaviors. However, even with MoCap data, controlling simulated humanoids remains … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 9 publications
0
0
0
Order By: Relevance
“…Distillation has been applied in previous efforts to scale up reinforcement learning to train multi-task control policies [Merel et al 2019[Merel et al , 2020. Wagener et al [2023] used a single-stage distillation approach to train a general motion tracking controller capable of imitating approximately 3.5 hours of motion data. Our multi-stage progress distillation approach enables our system to train versatile controllers on 8.5 hours of text-labeled motion clips, leading to a unified end-to-end controller that can be directed to perform a large variety skills with simple text commands.…”
Section: Language-directed Controllersmentioning
confidence: 99%
See 1 more Smart Citation
“…Distillation has been applied in previous efforts to scale up reinforcement learning to train multi-task control policies [Merel et al 2019[Merel et al , 2020. Wagener et al [2023] used a single-stage distillation approach to train a general motion tracking controller capable of imitating approximately 3.5 hours of motion data. Our multi-stage progress distillation approach enables our system to train versatile controllers on 8.5 hours of text-labeled motion clips, leading to a unified end-to-end controller that can be directed to perform a large variety skills with simple text commands.…”
Section: Language-directed Controllersmentioning
confidence: 99%
“…Our approach is inspired by the first stage of MoCapAct [Wagener et al 2023], leveraging DeepMimic as our tracking method [Peng et al 2018]. We train a DeepMimic expert policy πœ‹ 𝑒 𝑖 (a 𝑑 |o 𝑑 , πœ™) on every motion capture sequence in our dataset, conditioned on the current state of character o 𝑑 as well as a phase variable πœ™ ∈ [0, 1] that synchronizes the policy to the reference motion.…”
Section: Training Per-motion Expert Tracking Policiesmentioning
confidence: 99%