2016
DOI: 10.1038/srep19133
|View full text |Cite
|
Sign up to set email alerts
|

Temperature based Restricted Boltzmann Machines

Abstract: Restricted Boltzmann machines (RBMs), which apply graphical models to learning probability distribution over a set of inputs, have attracted much attention recently since being proposed as building blocks of multi-layer learning systems called deep belief networks (DBNs). Note that temperature is a key factor of the Boltzmann distribution that RBMs originate from. However, none of existing schemes have considered the impact of temperature in the graphical model of DBNs. In this work, we propose temperature bas… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
18
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 28 publications
(23 citation statements)
references
References 60 publications
3
18
0
Order By: Relevance
“…One can observe the best results were obtained by DBM when using T ∈ {0.1, 0.2, 0.5}. Also, DBN-CD benefit from lower temperatures, thus confirming the results obtained by Lin et al [16], i.e. the lower the temperature the higher the entropy.…”
Section: Resultssupporting
confidence: 86%
See 1 more Smart Citation
“…One can observe the best results were obtained by DBM when using T ∈ {0.1, 0.2, 0.5}. Also, DBN-CD benefit from lower temperatures, thus confirming the results obtained by Lin et al [16], i.e. the lower the temperature the higher the entropy.…”
Section: Resultssupporting
confidence: 86%
“…https://archive.ics.uci.edu/ml/datasets/Semeion+Handwritten+Digit 7. Similar architectures have been commonly employed in the literature[12,16,24,25,34] 8. Notice all parameters and architectures have been empirically chosen[21].…”
mentioning
confidence: 99%
“…Its impact is evaluated through the learning steps, and the results are compared even with distinct activation functions, once such parameter added to the energy function can be interpreted as a scalar multiplication of the Sigmoid function input. Provided results confirm the hypothesis suggested by Li et al [13] that lower temperatures tend to reach more accurate results, as presented in Table I. Furthermore, one can observe that lower temperatures also support sparseness representations of the hidden layer, which leads to a dropout like regularization.…”
Section: Temperature-based Deep Boltzmann Machinessupporting
confidence: 89%
“…The even or odd sector of momentum values correspond to whether periodic or antiperiodic boundary conditions are imposed on the free fermion operator. We interpret the input data, vectors |i of 256 grayscale values for each pixel, as the eigenstates of the Hamiltonian (21). In order to do that, we binarize the MNIST data by setting a pixel value to 0 if it is smaller than 256/2, and 1 otherwise.…”
Section: A Simple Gge Machine For the Mnist Datasetmentioning
confidence: 99%
“…While there is a computational cost associated to calculating the charges we feed to the network, this is still a decrease in the total cost. The key difference here is the fact that the GGE algorithm assumes a simple Hamiltonian (21) with homogeneous coupling, whereas the RBM learns an inhomogeneous Hamiltonian with many different coupling constants.…”
Section: The Algorithm and Performance On The Mnist Datasetmentioning
confidence: 99%