2018
DOI: 10.1007/978-3-030-01228-1_17
|View full text |Cite
|
Sign up to set email alerts
|

MT-VAE: Learning Motion Transformations to Generate Multimodal Human Dynamics

Abstract: Long-term human motion can be represented as a series of motion modes-motion sequences that capture short-term temporal dynamics-with transitions between them. We leverage this structure and present a novel Motion Transformation Variational Auto-Encoders (MT-VAE) for learning motion sequence generation. Our model jointly learns a feature embedding for motion modes (that the motion sequence can be reconstructed from) and a feature transformation that represents the transition of one motion mode to the next moti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
99
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 111 publications
(102 citation statements)
references
References 47 publications
0
99
0
Order By: Relevance
“…It enables realistic joint rotations transformation across subjects through cycle consistency based adversarial training, and then the subject-specific motion is generated by the prior-known basic skeleton and the kinematics. Yan et al [38] adopted a variational recurrent auto-encoder to perform the multi-modal motion transformation from one to another. The embedding feature was generated by the variational auto-encoder and it was further enhanced by motion retargeting.…”
Section: B Cross-subject Human Motion Retargetingmentioning
confidence: 99%
“…It enables realistic joint rotations transformation across subjects through cycle consistency based adversarial training, and then the subject-specific motion is generated by the prior-known basic skeleton and the kinematics. Yan et al [38] adopted a variational recurrent auto-encoder to perform the multi-modal motion transformation from one to another. The embedding feature was generated by the variational auto-encoder and it was further enhanced by motion retargeting.…”
Section: B Cross-subject Human Motion Retargetingmentioning
confidence: 99%
“…RNNs have been extensively explored for short and longterm human motion prediction in computer vision [Fragkiadaki et al 2015;Martinez et al 2017]. It is commonly noted that these models can often be unstable for long-term sequence prediction, and the production of long-term stable sequences is considered an accomplishment, even in the absence of a control task [Habibie et al 2017;Yan et al 2018;Zhou et al 2018].…”
Section: Kinematic Motion Synthesismentioning
confidence: 99%
“…In [3], a pose generator and motion generator are trained progressively with the help of GANs for human 2D motion generation. Other methods [13,36] directly generate 2D human motions by VAEs or GANs. However, direct modeling of motions in 2D is inherently insufficient to capture the underlying 3D human shape articulations.…”
Section: Introductionmentioning
confidence: 99%
“…However, direct modeling of motions in 2D is inherently insufficient to capture the underlying 3D human shape articulations. Existing methods [3,18,36,37] often approach human poses in terms of the joints' coordinate locations, which unnecessarily entangle the human skeletons and their motion trajectories, and introduce extra barriers in faithful modeling of human kinematics. To further complicate the matter, the variation of generated human dynamics could be severely limited by the initial priors [37].…”
Section: Introductionmentioning
confidence: 99%