2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.241
|View full text |Cite
|
Sign up to set email alerts
|

Learning from Simulated and Unsupervised Images through Adversarial Training

Abstract: With recent progress in graphics, it has become more tractable to train models on synthetic images, potentially avoiding the need for expensive annotations. However, learning from synthetic images may not achieve the desired performance due to a gap between synthetic and real image distributions. To reduce this gap, we propose Simulated+Unsupervised (S+U) learning, where the task is to learn a model to improve the realism of a simulator's output using unlabeled real data, while preserving the annotation inform… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
1,198
0
2

Year Published

2017
2017
2019
2019

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 1,591 publications
(1,201 citation statements)
references
References 47 publications
1
1,198
0
2
Order By: Relevance
“…Typically used fine-tuning methods require prohibitively large labeled data in the target domain. As an alternative, domain adaptation methods attempt to minimize domain shift either by feature sharing[2] or by learning to reconstruct the target from source domain[3, 4]. In essence, domain adaptation methods learn the marginal distributions [5] to transform source to target domain.…”
Section: Introductionmentioning
confidence: 99%
“…Typically used fine-tuning methods require prohibitively large labeled data in the target domain. As an alternative, domain adaptation methods attempt to minimize domain shift either by feature sharing[2] or by learning to reconstruct the target from source domain[3, 4]. In essence, domain adaptation methods learn the marginal distributions [5] to transform source to target domain.…”
Section: Introductionmentioning
confidence: 99%
“…This can be handled by using paired source-target data to regularize the translation [9]. In the unpaired setting, prior work has constrained the translated images to be close to the source images [30] but this only works for small domain shifts. Most current methods use a combination of task-specific losses (i.e., preserving the task network's output after translation) [25], image-space and feature-space adversarial losses, cycleconsistency losses, and semantic losses (i.e., preserving the semantics of the image after translation) [14,26,42].…”
Section: Related Workmentioning
confidence: 99%
“…While our focus is on improving visual quality, our method can be used for domain adaptation. Deep networks can be trained on large-scale, labeled synthetic datasets [28,41] and prior work has looked at adapting them to improve their performance on real data [9,30,25]. Many of these methods impose a task-specific loss on this adaptation [14,26,42].…”
Section: Introductionmentioning
confidence: 99%
“…While the number of biological samples might be limited, realistic in silico generation of observations could accommodate for this unfavorable setting. In practice, in silico generation has seen success in computer vision when used for 'data augmentation', whereby in silico generated samples are used alongside the original ones to artificially increase the number of observations 3 . While classically realistic data modeling relies on a thorough understanding of the laws underlying the production of such data, current methods of choice for photo-realistic image generation rely on Deep Learning-based Generative Adversarial Networks (GANs) [4][5][6][7] and Variational Autoencoders (VAEs) 8,9 .…”
Section: Introductionmentioning
confidence: 99%