2017
DOI: 10.48550/arxiv.1710.05050
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning Independent Features with Adversarial Nets for Non-linear ICA

Abstract: Reliable measures of statistical dependence could be useful tools for learning independent features and performing tasks like source separation using Independent Component Analysis (ICA). Unfortunately, many of such measures, like the mutual information, are hard to estimate and optimize directly. We propose to learn independent features with adversarial objectives (Goodfellow et al., 2014;Huszar, 2016) which optimize such measures implicitly. These objectives compare samples from the joint distribution and th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
45
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 26 publications
(45 citation statements)
references
References 12 publications
0
45
0
Order By: Relevance
“…However, both of these bounds require a large number of negative samples and as a result, recent methods either train with extremely large batch-sizes [7,40] or an additional memory-bank of negative samples [8,48]. Unlike these works, we estimate mutual information using a Jensen-Shannon Divergence (JSD) bound, similar to formulations used for generative modeling [36] and source separation [5]. This bound on mutual information is derived by replacing the KL-divergence in equation 2 with the Jensen-Shannon divergence.…”
Section: Mutual Information Maximizationmentioning
confidence: 99%
“…However, both of these bounds require a large number of negative samples and as a result, recent methods either train with extremely large batch-sizes [7,40] or an additional memory-bank of negative samples [8,48]. Unlike these works, we estimate mutual information using a Jensen-Shannon Divergence (JSD) bound, similar to formulations used for generative modeling [36] and source separation [5]. This bound on mutual information is derived by replacing the KL-divergence in equation 2 with the Jensen-Shannon divergence.…”
Section: Mutual Information Maximizationmentioning
confidence: 99%
“…Several methods for learning disentangled representations have been recently proposed, however lacking identifiability guarantees, most of which are based on a variational autoencoder (VAE, Kingma and Welling [33]) formulation, and decompositions of the VAE objective, for example [19,31,8,7,35,6,14]. GAN-based approaches for disentanglement have been proposed as well [8,5].…”
Section: Related Workmentioning
confidence: 99%
“…We then generate the observed data via linear mixing. We consider one of the latent components as the condition and train the autoencoder to reconstruct the observed data, while obtaining code which is independent of the condition using the regression objective (5).…”
Section: D Analysis Demonstrationmentioning
confidence: 99%
“…Different from above methods, in our approach we use Jensen-Shannon divergence based mutual information estimation, similar to the formulations in (Nowozin, Cseke, and Tomioka 2016) and (Brakel and Bengio 2017),…”
Section: Mutual Information Maximizationmentioning
confidence: 99%