2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.00159
|View full text |Cite
|
Sign up to set email alerts
|

High-Fidelity Neural Human Motion Transfer from Monocular Video

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 24 publications
(9 citation statements)
references
References 26 publications
0
9
0
Order By: Relevance
“…Many approaches formulate the motion transfer problem as an image-to-image translation task. Kappel et al [15] divided the image translation task into four cascaded generative networks and proposed a structure network to learn wrinkles of garments, which generates high-quality results. Zhang et al [49] proposed a decoupled GAN to disentangle the shape and texture of clothing.…”
Section: Novel View/pose Synthesismentioning
confidence: 99%
See 1 more Smart Citation
“…Many approaches formulate the motion transfer problem as an image-to-image translation task. Kappel et al [15] divided the image translation task into four cascaded generative networks and proposed a structure network to learn wrinkles of garments, which generates high-quality results. Zhang et al [49] proposed a decoupled GAN to disentangle the shape and texture of clothing.…”
Section: Novel View/pose Synthesismentioning
confidence: 99%
“…With the corresponding UV-map, the geometry is rasterized using a neural texture by bilinear sampling and then is translated to an RGB image using a neural network. We compare our method with three state-of-the-art methods Neural Body [33], HF-NHMT [15] and StylePeople [12]. The trained models of [33] and [15] are generated by the official implementations, and the trained models of [12] on 20 videos of SelfieVideo are provided by the authors.…”
Section: Applicationsmentioning
confidence: 99%
“…Head puppetry or "talking head generation" is the task of generating a plausible video of a talking head from a source image or video by mimicking the movements and facial expressions of a reference video (Zakharov et al [2019]), while lip syncing consists in synchronizing lip movements on a video to match a target speech segment (Prajwal et al [2020]). Head puppetry and lip syncing are both forms of motion transfer, which refers more broadly to the task of mapping the motion of a given individual in source video to the motion of another individual in a target image or video (Zhu et al [2021], Kappel et al [2021]). Face swapping, head puppetry, and lip syncing are commonly referred to as "deepfakes" because they can be used to usurp someone's identity in a video; however, they involve distinct generation pipelines.…”
Section: Local Partially Synthetic Dlsammentioning
confidence: 99%
“…[11,34] or videos of real people e.g. [23,14,20]. Some works explored transferring the style between rigid 3D objects.…”
Section: Related Workmentioning
confidence: 99%