2020
DOI: 10.1109/jsait.2020.2983071
|View full text |Cite
|
Sign up to set email alerts
|

PacGAN: The Power of Two Samples in Generative Adversarial Networks

Abstract: Generative adversarial networks (GANs) are innovative techniques for learning generative models of complex data distributions from samples. Despite remarkable recent improvements in generating realistic images, one of their major shortcomings is the fact that in practice, they tend to produce samples with little diversity, even when trained on diverse datasets. This phenomenon, known as mode collapse, has been the main focus of several recent advances in GANs. Yet there is little understanding of why mode coll… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

4
189
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 147 publications
(208 citation statements)
references
References 41 publications
4
189
0
Order By: Relevance
“…In Fig. 4, we qualitatively demonstrate the results on synthetic point sets where our method achieve comparable or even better results over VEEGAN [39] and PacGAN [26] without using two or more real samples at a time.…”
Section: Synthetic Datasetsmentioning
confidence: 73%
“…In Fig. 4, we qualitatively demonstrate the results on synthetic point sets where our method achieve comparable or even better results over VEEGAN [39] and PacGAN [26] without using two or more real samples at a time.…”
Section: Synthetic Datasetsmentioning
confidence: 73%
“…To avoid the problem, we exploit the recent PacGAN [16] technique: it consists in passing a set of samples to the discrimination function instead of a single one. PacGAN is intended to tackle the mode collapse problem in GAN training.…”
Section: Proposed Approachmentioning
confidence: 99%
“…The underlying principle being that if a set of images are sampled from the same training set, they are very likely to be completely different, whereas if the generator experiences mode collapse, generated images are likely to be similar. In practice, we only give two samples to the discriminator, which is sufficient to overcome the loss of diversity as suggested in [16]. The resulting training procedure is summarized in Algorithm 2.…”
Section: Proposed Approachmentioning
confidence: 99%
“…For example, CIFAR-10 has 50,000 training images and, assuming a batch size of 64, one epoch represents roughly 781 iterations for the generator. 4 σ(y) = 1/1 + e −y Modes (max 1000) KL DCGAN (Radford, Metz, and Chintala, 2015) 99.0 3.40 ALI (Dumoulin et al, 2016) 16.0 5.40 Unrolled GAN (Metz et al, 2016) 48.7 4.32 VEEGAN (Srivastava et al, 2017a) 150.0 2.95 PacGAN (Lin et al, 2017) 1000.0 ± 0.00 0.06 ± 1.0e −2 GAN+MINE (Belghazi et al, 2018) 1000.0 ± 0.00 0.05 ± 6.3e −3 acGAN (3D)…”
Section: Performance Of Acgan Against Existing Baselinesmentioning
confidence: 99%
“…We report our results (averaged over 10 runs) in Table 2 and compare them with other existing baselines in the literature. Our method could recover all 1000 modes like PaCGAN (Lin et al, 2017) and MINE (Belghazi et al, 2018); these two approaches either increase the dimensionality of the generator inputs either by packing multiple samples or by adding a latent code vector which helps overcoming mode collapse. Generated samples are shown in Fig.…”
Section: Performance Of Acgan Against Existing Baselinesmentioning
confidence: 99%