2017 IEEE 27th International Workshop on Machine Learning for Signal Processing (MLSP) 2017
DOI: 10.1109/mlsp.2017.8168140
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial nets with perceptual losses for text-to-image synthesis

Abstract: Recent approaches in generative adversarial networks (GANs) can automatically synthesize realistic images from descriptive text. Despite the overall fair quality, the generated images often expose visible flaws that lack structural definition for an object of interest. In this paper, we aim to extend state of the art for GAN-based text-to-image synthesis by improving perceptual quality of generated images. Differentiated from previous work, our synthetic image generator optimizes on perceptual loss functions t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 33 publications
(31 citation statements)
references
References 9 publications
0
31
0
Order By: Relevance
“…We can compare with it for a general measure of the image diversity. Following the procedure of Prog.GAN, we randomly sample ∼10, 000 image pairs from all generated samples 3 Table 3 right. HDGAN outperforms both methods.…”
Section: Comparative Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…We can compare with it for a general measure of the image diversity. Following the procedure of Prog.GAN, we randomly sample ∼10, 000 image pairs from all generated samples 3 Table 3 right. HDGAN outperforms both methods.…”
Section: Comparative Resultsmentioning
confidence: 99%
“…Recently, Dong et al [8] propose to learn a joint embedding of images and text so as to re-render a prototype image conditioned on a targeting description. Cha et al [3] explore the usage of the perceptional loss [16] with a CNN pretrained on Ima-geNet and Dash et al [6] make use of auxiliary classifiers (similar with [31]) to assist GAN training for text-to-image synthesis. Xu et al [43] shows an attention-driven method to improve fine-grained details.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Finally, we additionally impose a reconstruction loss L rec that encourages the predicted instance masks to be similar to the ground-truths. We implement this idea using perceptual loss [11,3,33,2], which measures the distance of real and fake images in the feature space of a pre-trained CNN by…”
Section: Shape Generationmentioning
confidence: 99%
“…In this work we introduce AR-GAN that differs from the previous approaches by optimizing on an activation reconstruction loss (Johnson et al, 2016;Cha et al 2017) in addition to regularizing the original GAN objective function and cycle-consistency optimizations to present visually more compelling synthetic images on an unaligned dataset. The main focus of this work is to analyze the performance of plant disease recognition systems using synthetically generated image data.…”
Section: Introductionmentioning
confidence: 99%