2014
DOI: 10.48550/arxiv.1411.7783
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

From neural PCA to deep unsupervised learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(19 citation statements)
references
References 0 publications
0
19
0
Order By: Relevance
“…Considering the bottom-up inference procedure and topdown generation procedure together, additional memory mechanisms can help to reduce the competition between invariant feature extraction and local variant reconstruction, especially when label information is provided (e.g., in supervised or semi-supervised setting). Similar idea is highlighted in the Ladder Network (Valpola, 2014;Rasmus et al, 2015), which reconstructs the input hierarchically using an extension of denoising autoencoders (dAEs) (Vincent et al, 2010) with the help of lateral connections and achieves excellent performance on semi-supervised learning (Rasmus et al, 2015). Though it is possible to interpret the Ladder Network probabilistically as in (Bengio et al, 2013b), we model the data likelihood directly with the help of external memory instead of explicit lateral edges.…”
Section: Related Workmentioning
confidence: 90%
See 2 more Smart Citations
“…Considering the bottom-up inference procedure and topdown generation procedure together, additional memory mechanisms can help to reduce the competition between invariant feature extraction and local variant reconstruction, especially when label information is provided (e.g., in supervised or semi-supervised setting). Similar idea is highlighted in the Ladder Network (Valpola, 2014;Rasmus et al, 2015), which reconstructs the input hierarchically using an extension of denoising autoencoders (dAEs) (Vincent et al, 2010) with the help of lateral connections and achieves excellent performance on semi-supervised learning (Rasmus et al, 2015). Though it is possible to interpret the Ladder Network probabilistically as in (Bengio et al, 2013b), we model the data likelihood directly with the help of external memory instead of explicit lateral edges.…”
Section: Related Workmentioning
confidence: 90%
“…We optimize the objective with a stochastic gradient variational Bayes (SGVB) method . Note that we cannot send the message of a intermediate layer in the recognition model to a layer in the generative model through a lateral connection as in Ladder Network (Valpola, 2014;Rasmus et al, 2015) because that indeed changes the distribution of p(x|z) according to the data x. However, we do not use any information of x in the generative model explicitly and the correctness of the variational bound can be verified.…”
Section: Q-net P-netmentioning
confidence: 99%
See 1 more Smart Citation
“…Our approach is based on the recently proposed Ladder Network training procedure that has proven to be very successful, especially in cases where the training dataset is composed of both labeled and unlabeled data [9] [10]. In addition to the supervised objective, the Ladder Network also has an unsupervised objective corresponding to the reconstruction costs of a stack of denoising autoencoders.…”
Section: Introductionmentioning
confidence: 99%
“…3.2]. We note that the adopted architecture is similar to the Ladder network [23], where the clean pathway is used for prediction while the corrupted one guaranties that the network is noise-invariant. As done in [4], and in order to avoid over-fitting, we add a reconstruction loss function to our objectives MI-ADM in Eq.…”
mentioning
confidence: 99%