2022
DOI: 10.48550/arxiv.2204.07172
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Diagnosing and Fixing Manifold Overfitting in Deep Generative Models

Abstract: Likelihood-based, or explicit, deep generative models use neural networks to construct flexible high-dimensional densities. This formulation directly contradicts the manifold hypothesis, which states that observed data lies on a low-dimensional manifold embedded in high-dimensional ambient space. In this paper we investigate the pathologies of maximum-likelihood training in the presence of this dimensionality mismatch. We formally prove that degenerate optima are achieved wherein the manifold itself is learned… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 17 publications
0
3
0
Order By: Relevance
“…To obtain such samples, we follow the two-stage procedure described in [39][40][41], where samples from the lower-dimensional q ψ (z) are obtained through a score-based generative model. These models have shown tremendous performance in fitting complex distributions [10,42], an ability which aligns with our objective of learning the distribution within a multimodal latent space.…”
Section: Our Approach: Multimodal Latent Diffusionmentioning
confidence: 99%
See 1 more Smart Citation
“…To obtain such samples, we follow the two-stage procedure described in [39][40][41], where samples from the lower-dimensional q ψ (z) are obtained through a score-based generative model. These models have shown tremendous performance in fitting complex distributions [10,42], an ability which aligns with our objective of learning the distribution within a multimodal latent space.…”
Section: Our Approach: Multimodal Latent Diffusionmentioning
confidence: 99%
“…Joint generation: To generate a new sample for all modalities, we use a simple scorebased diffusion model in latent space [32,39,40,42,43]. This requires reversing a stochastic noising process, starting from a simple Gaussian distribution.…”
Section: Joint and Conditional Multimodal Latent Diffusion Processesmentioning
confidence: 99%
“…Similarly, [76] and [5] use a deterministic autoencoder to learn the latent representation of the data, and a normalizing flow to model the distribution of such latents, leading to better density estimation while avoiding over-regularization of the latent variables. More recently, [46] discusses the problem of manifold overfitting, which arises when the manifold is learned but not the distribution on it, and propose a two-step training procedure applicable to all the likelihood-based models. NFs on manifolds: Several authors propose variations of NFs that can model data on a manifold.…”
Section: Related Workmentioning
confidence: 99%