2021
DOI: 10.48550/arxiv.2106.09620
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Disentangling Identifiable Features from Noisy Data with Structured Nonlinear ICA

Abstract: We introduce a new general identifiable framework for principled disentanglement referred to as Structured Nonlinear Independent Component Analysis (SNICA). Our contribution is to extend the identifiability theory of deep generative models for a very broad class of structured models. While previous works have shown identifiability for specific classes of time-series models, our theorems extend this to more general temporal structures as well as to models with more complex structures such as spatial dependencie… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 24 publications
0
4
0
Order By: Relevance
“…More recently, Khemakhem et al (2020a) proved a major breakthrough by showing that given side information u, identifiability of the entire generative model is possible up to certain (nonlinear) equivalences. Since this pathbreaking work, many generalizations have been proposed (Hälvä and Hyvarinen, 2020;Hälvä et al, 2021;Khemakhem et al, 2020b;Li et al, 2019;Mita et al, 2021;Sorrenson et al, 2019;Yang et al, 2021;Klindt et al, 2020;Brehmer et al, 2022), all of which require some form of auxiliary information. Other approaches to identifiability include various forms of weak supervision such as contrastive learning (Zimmermann et al, 2021), group-based disentanglement (Locatello et al, 2020), and independent mechanisms (Gresele et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…More recently, Khemakhem et al (2020a) proved a major breakthrough by showing that given side information u, identifiability of the entire generative model is possible up to certain (nonlinear) equivalences. Since this pathbreaking work, many generalizations have been proposed (Hälvä and Hyvarinen, 2020;Hälvä et al, 2021;Khemakhem et al, 2020b;Li et al, 2019;Mita et al, 2021;Sorrenson et al, 2019;Yang et al, 2021;Klindt et al, 2020;Brehmer et al, 2022), all of which require some form of auxiliary information. Other approaches to identifiability include various forms of weak supervision such as contrastive learning (Zimmermann et al, 2021), group-based disentanglement (Locatello et al, 2020), and independent mechanisms (Gresele et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…This contrasts a recent line of work that has established fundamental new results regarding the identifiability of VAEs that requires conditioning on an auxiliary variable u that renders each latent dimension conditionally independent (Khemakhem et al, 2020a). While this result has been generalized and relaxed in several directions (Hälvä and Hyvarinen, 2020;Hälvä et al, 2021;Khemakhem et al, 2020b;Li et al, 2019;Mita et al, 2021;Sorrenson et al, 2019;Yang et al, 2021;Klindt et al, 2020;Brehmer et al, 2022), fundamentally these results still crucially rely on the side information u. We show that this is in fact unnecessary-confirming existing empirical studies (e.g Willetts and Paige, 2021;Falck et al, 2021)-and do so without sacrificing any representational capacity.…”
Section: Introductionmentioning
confidence: 99%
“…For time-series data, history information is widely used as side information for nonlinear ICA. However, most existing work that establishes identifiability results considers either stationary independent sources such as PCL (Hyvarinen & Morioka, 2017), SlowVAE (Klindt et al, 2020) or under lin-ear transition assumptions such as SlowVAE (Klindt et al, 2020) and SNICA (Hälvä et al, 2021), or with certain structure such as Markov properties in HM-NLICA (Hälvä & Hyvarinen, 2020). LEAP (Yao et al, 2021), which is the closest work to ours, has established the identifiability of the nonparametric latent temporal processes in certain nonstationary cases, under the condition that the distribution of the noise terms of the latent processes vary across segments.…”
Section: Introductionmentioning
confidence: 99%
“…These methods typically require auxiliary information or weak supervision for identifiability. Examples of weak supervision include the following strategies: using auxiliary information (Khemakhem et al, 2020;Locatello et al, 2019b); using paired samples (Locatello et al, 2020); using data augmentation (von Kügelgen et al, 2021); and leveraging temporal or spatial dependencies among the samples (Hälvä et al, 2021). In contrast, the Sparse VAE does not need auxiliary information or weak supervision for identifiability; instead, the anchor feature assumption is sufficient.…”
Section: Introductionmentioning
confidence: 99%