2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.00511
|View full text |Cite
|
Sign up to set email alerts
|

StylePeople: A Generative Model of Fullbody Human Avatars

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
41
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 57 publications
(43 citation statements)
references
References 37 publications
0
41
0
Order By: Relevance
“…1). We show that our learned representation handles occlusions more effectively than other techniques [66,52,54,11] that just take geometry priors (e.g., UV, depth or normal maps) from a coarse mesh as input (see Fig. 3, 7).…”
Section: Related Workmentioning
confidence: 92%
See 1 more Smart Citation
“…1). We show that our learned representation handles occlusions more effectively than other techniques [66,52,54,11] that just take geometry priors (e.g., UV, depth or normal maps) from a coarse mesh as input (see Fig. 3, 7).…”
Section: Related Workmentioning
confidence: 92%
“…A 2D convolutional network is often utilized for both shape completion and appearance synthesis in one stage [66,52]. However, [66,52,54,11] do not reconstruct geometry explicitly and cannot handle self-occlusions effectively. In contrast, our rendering is conditioned on a learned 3D volumetric representation using a two-stage approach (see Fig.…”
Section: Related Workmentioning
confidence: 99%
“…In human generation research, most of the existing applications focus on precise control of pose and appearance by leveraging conditional VAE and U-Net [17,73] or StyleGAN-related architectures [3,24,44,72]. Specifically, the 3D method [24] renders StyleGAN-generated neural textures on the parametric human models, but the results are restricted by the quantity and quality of training data. The other works [3,72,73] preserve texture quality by spatial modulation using the extracted UV texture map, and perform pose transfer conditioned by extracted pose features.…”
Section: Human Generationmentioning
confidence: 99%
“…2D and 3D Generative Models: Most modern methods for synthesizing natural images leverage generative adversarial networks (GANs) [15] or variational auto-encoders (VAEs) [24]. These methods have achieved a high level of photorealism [21][22][23] and can yield impressive results on the task of synthesizing 2D images of humans [4,16,26,31,50]. However, such methods reason in 2D and hence 3D consistency cannot be guaranteed [26,31] nor is extracting 3D geometry from such approaches straightforward.…”
Section: Related Workmentioning
confidence: 99%