2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01092
|View full text |Cite
|
Sign up to set email alerts
|

RiFeGAN: Rich Feature Generation for Text-to-Image Synthesis From Prior Knowledge

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
42
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 93 publications
(42 citation statements)
references
References 16 publications
0
42
0
Order By: Relevance
“…GAN-based Text-to-image generation. In the past few years, Generative Adversarial Networks (GANs) [18] have shown promising results on text-to-image generation [5,8,9,14,17,22,[27][28][29][30]34,39,40,46,47,54,[56][57][58][62][63][64][65][66][67][68]. GAN-INT-CLS [46] was the first to use a conditional GAN formulation for text-to-image generation.…”
Section: Related Workmentioning
confidence: 99%
“…GAN-based Text-to-image generation. In the past few years, Generative Adversarial Networks (GANs) [18] have shown promising results on text-to-image generation [5,8,9,14,17,22,[27][28][29][30]34,39,40,46,47,54,[56][57][58][62][63][64][65][66][67][68]. GAN-INT-CLS [46] was the first to use a conditional GAN formulation for text-to-image generation.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, exploring both the textual and visual representations has been studied in many challenging tasks. In the image domain, some works have dealt with image caption [43], image grounding [21], and text-to-image synthesis [6]. In the video domain, some works focus on temporal localization using natural language [32,35,51], where the temporal boundary needs to be localized with a given natural language description.…”
Section: Textual and Visual Understandingmentioning
confidence: 99%
“…We note however, that the FID score (which is reference-based and compares the distributions of real and synthetic images together) has been observed to be more consistent with human judgement of image realism than IS (which is reference-free and does not make comparisons to real images) (Heusel et al, 2017). We were not able to recompute other metrics for RiFeGAN (Cheng et al, 2020) and LeciaGAN (Qiao et al, 2019) as the pretrained models have not been made publicly available. In Table 5, "-" represents cases where the data was not reported or is reported in a manner which is non-comparable (besides FID values).…”
Section: A4 Note On Evaluation Metricsmentioning
confidence: 99%