2014
DOI: 10.1016/j.patcog.2013.05.025
|View full text |Cite
|
Sign up to set email alerts
|

Training restricted Boltzmann machines: An introduction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

5
205
0
3

Year Published

2016
2016
2023
2023

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 405 publications
(213 citation statements)
references
References 20 publications
5
205
0
3
Order By: Relevance
“…Based on the feature that the DBN only learns using the data without labels, two types of data sets are present in this flow chart, namely, training samples without labels and labeled samples for fine tuning [31]. The training samples are used to train the DBN and the SOM; the fine tuning samples are a subset of training samples that have been manually labeled [26,27,32].…”
Section: Network Trainingmentioning
confidence: 99%
“…Based on the feature that the DBN only learns using the data without labels, two types of data sets are present in this flow chart, namely, training samples without labels and labeled samples for fine tuning [31]. The training samples are used to train the DBN and the SOM; the fine tuning samples are a subset of training samples that have been manually labeled [26,27,32].…”
Section: Network Trainingmentioning
confidence: 99%
“…We refer the interested reader to Bengio (2009), Hinton (2012, and Fischer & Igel (2014) for detailed derivations. Figure 1 shows a schematic of the the deep-belief network.…”
Section: Robertmentioning
confidence: 99%
“…Figure 1 shows a schematic of the the deep-belief network. The multi-layer DBN can be constructed from several Restricted Boltzmann Machines (Freund & Haussler 1992;Bishop 2006;Le Roux & Bengio 2008;Bengio 2009Bengio , 2012Lee et al 2011a;Hinton 2012;Montavon et al 2012;Fischer & Igel 2014) with the addition of a logistic regression layer at the top of the network. The RBM is a two-layer neural network able to learn the underlying probability distribution over its set of input values.…”
Section: Robertmentioning
confidence: 99%
“…We update the parameters of the subspaceRBM using contrastive divergence learning procedure [8,10]. For this purpose, we need to calculate the gradient of the log-likelihood function.…”
Section: Learningmentioning
confidence: 99%