2021
DOI: 10.48550/arxiv.2110.10804
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Identifiable Deep Generative Models via Sparse Decoding

Abstract: We develop the Sparse VAE, a deep generative model for unsupervised representation learning on high-dimensional data. Given a dataset of observations, the Sparse VAE learns a set of latent factors that captures its distribution. The model is sparse in the sense that each feature of the dataset (i.e., each dimension) depends on a small subset of the latent factors. As examples, in ratings data each movie is only described by a few genres; in text data each word is only applicable to a few topics; in genomics, e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 14 publications
0
6
0
Order By: Relevance
“…These include Laplacian [2], Spike-and-Slab [52], and Beta-Bernoulli [50] distributions. Other work has incorporated sparsity through hierarchical posterior distributions [47] or through group-sparsity in connections to generator networks [1,34]. We refer the reader to Zhang et al [58] for a review of variational inference.…”
Section: Related Workmentioning
confidence: 99%
“…These include Laplacian [2], Spike-and-Slab [52], and Beta-Bernoulli [50] distributions. Other work has incorporated sparsity through hierarchical posterior distributions [47] or through group-sparsity in connections to generator networks [1,34]. We refer the reader to Zhang et al [58] for a review of variational inference.…”
Section: Related Workmentioning
confidence: 99%
“…Our approach is to avoid additional forms of supervision altogether, and enforce identifiability in a purely unsupervised fashion. Recent work along these lines includes , who propose to use Brenier maps and input convex neural networks, and Moran et al (2021) who leverage sparsity and an anchor feature assumption. Aside from different assumptions, the main difference between this line of work and our work is that their work only identifies the latent space P (Z), whereas our focus is on jointly identifying both P (Z) and f .…”
Section: Related Workmentioning
confidence: 99%
“…To the best of our knowledge, our results are the first to establish identifiability of both the latent space and decoder for deep generative models without conditioning in the latent space or weak supervision. We note that and Moran et al (2021) also propose deep architectures that identify the latent space, but not the decoder.…”
Section: Generative Modelmentioning
confidence: 99%
“…Importantly, as many functions can perfectly fit the training data-i.e. the function represented by the DNN model is often nonidentifiable (Raue et al, 2013;Martín & González, 2010;Moran et al, 2021), and given an input x the uncertainty estimate incorporates quantification if possible hypothesis function "agree" on the prediction of x. It is thus natural to consider that training with perturbations that maximize this quantity will most efficiently narrow down the hypothesis set, leaving out the functions that increase the margin.…”
Section: Introductionmentioning
confidence: 99%