2019
DOI: 10.1007/978-3-030-11479-4_13
|View full text |Cite
|
Sign up to set email alerts
|

Why Dose Layer-by-Layer Pre-training Improve Deep Neural Networks Learning?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 27 publications
0
3
0
Order By: Relevance
“…Specifically, by adding a sigmoid layer on top of a DBN and reusing the generatively trained weights as the initial weights, we can discriminatively train the underlying MLP (Bengio, 2007) via conventional back-propagation-based techniques to converge to a more accurate local optimum. Pre-training differentiates itself from the SSL techniques by finding a proper initial point within the complex search space in an informed way, without modifying the objective function (Erhan, 2010).…”
Section: Deep Belief Networkmentioning
confidence: 99%
“…Specifically, by adding a sigmoid layer on top of a DBN and reusing the generatively trained weights as the initial weights, we can discriminatively train the underlying MLP (Bengio, 2007) via conventional back-propagation-based techniques to converge to a more accurate local optimum. Pre-training differentiates itself from the SSL techniques by finding a proper initial point within the complex search space in an informed way, without modifying the objective function (Erhan, 2010).…”
Section: Deep Belief Networkmentioning
confidence: 99%
“…Pre-training methods are used to find the initial values of network weights and free the learning process from the local minimums in the middle of the road as a fundamental obstacle in the training process. These methods seek to find an appropriate starting point for network weights and, in addition to facilitating the network training process, also improve the generalizability of the network [69]. In 2006, Hinton proposed the Restrict Boltzmann Machine (RBM) method for pretraining multilayer neural networks to reduce the nonlinear dimension [51].…”
Section: Pre-trainingmentioning
confidence: 99%
“…In 2015, Seyyed Salehi et al introduced the layer-by-layer pre-training method for pretraining Autoencoder Deep Bottleneck Networks to extract the principal components [50]. However, we used a bidirectional version of this method to pre-train DNNs [69]. This method is used to converge fully connected networks with neurons with sigmoid and sigmoid tangent nonlinearity.…”
Section: Pre-trainingmentioning
confidence: 99%