2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.01737
|View full text |Cite
|
Sign up to set email alerts
|

Towards Implicit Text-Guided 3D Shape Generation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
44
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 43 publications
(45 citation statements)
references
References 48 publications
1
44
0
Order By: Relevance
“…Relying on a pre-trained CLIP model, CLIPMesh directly performs optimization on mesh parameters to generate shape and texture that enables to render images matching the input text prompt. Instead of using CLIP, Liu et al [202] adopt a BERT-based text encoder and achieve text-guided 3D shape generation with high-fidelity.…”
Section: Other Methodsmentioning
confidence: 99%
“…Relying on a pre-trained CLIP model, CLIPMesh directly performs optimization on mesh parameters to generate shape and texture that enables to render images matching the input text prompt. Instead of using CLIP, Liu et al [202] adopt a BERT-based text encoder and achieve text-guided 3D shape generation with high-fidelity.…”
Section: Other Methodsmentioning
confidence: 99%
“…There are mainly two categories of methods, i.e., fully-supervised and optimizationbased methods. The fully-supervised method [6,8,25] uses ground truth text and the paired 3D objects with explicit 3D representations as training data. Specifically, CLIP-Forge [43] uses a two-stage training scheme, which consists of shape autoencoder training, and conditional normalizing flow training.…”
Section: Text-guided 3d Shape Generationmentioning
confidence: 99%
“…In Fig. 6, we first compare with the method TITG3SG [25], which is an encoder-decoder-based framework. They only experiment with object classes of Chair and Table . Here we conduct the comparison qualitatively.…”
Section: Qualitative Comparisonmentioning
confidence: 99%
See 1 more Smart Citation
“…Colorization can be formulated to this task and handled by Generative Adversarial Networks [11] (GAN) based approaches [19,41,35,30,44]. They employ an adversarial loss that learns to discriminate between real and generated images, and then minimize this loss by updating the generator to make the produced results look realistic [57,28,31,50,36,51,45,46,42].…”
Section: Gan-based Image-to-image Translationmentioning
confidence: 99%