2015
DOI: 10.48550/arxiv.1503.03167
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Convolutional Inverse Graphics Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
19
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(20 citation statements)
references
References 0 publications
0
19
0
Order By: Relevance
“…Researchers have encouraged invariance or forgetting with VAEs by utilizing various combinations of weak supervision [1], [23], adversarial training [6], and by trading off reconstruction error against encoding capacity according to the information bottleneck principle [25]. Others have sought to fully disentangle the latent space into independent partitions [14], thereby enabling the isolation of wanted from unwanted information. Whilst unsupervised methods for disentanglement exist [4], [5], recent research indicates that such methods do not achieve disentanglement consistently [19] and, even if they do, they do not allocate factors to the latent space in a predictable way.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Researchers have encouraged invariance or forgetting with VAEs by utilizing various combinations of weak supervision [1], [23], adversarial training [6], and by trading off reconstruction error against encoding capacity according to the information bottleneck principle [25]. Others have sought to fully disentangle the latent space into independent partitions [14], thereby enabling the isolation of wanted from unwanted information. Whilst unsupervised methods for disentanglement exist [4], [5], recent research indicates that such methods do not achieve disentanglement consistently [19] and, even if they do, they do not allocate factors to the latent space in a predictable way.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Kulkarni et al [15] present an inverse graphics model that disentangles pose and lighting. They propose a novel training procedure for factoring their latent space.…”
Section: Related Workmentioning
confidence: 99%
“…Many equivalent parametrisations of the manifold may be possible, some allowing a more semantic interpretation of each latent coefficient than others, and the standard AE or DAE does not distinguish among these parametrisations. There has been recent interest in adapting the training schemes of autoencoders such that the latent representation is explicitly semantic, capturing attributes of the input in specific subsets of the latent variables [3,4]. We refer to these as partitioned autoencoders since the latent variables are partitioned into subsets which are treated differently from each other during training.…”
Section: Partitioned Autoencodersmentioning
confidence: 99%
“…We refer to these as partitioned autoencoders since the latent variables are partitioned into subsets which are treated differently from each other during training. Crucially, in this prior work the training scheme relies heavily on the existence of large structured datasets: in [4] a balanced dataset of labelled digit images; in [3] a dataset of faces constructed through systematic variation of attributes such as pose and lighting. Without such known structure in the training data, their proposed training schemes will be either impossible to apply, or biased by the presence of unbalanced or correlated factors in the training data.…”
Section: Partitioned Autoencodersmentioning
confidence: 99%