2016
DOI: 10.1162/neco_a_00848
|View full text |Cite
|
Sign up to set email alerts
|

An Infinite Restricted Boltzmann Machine

Abstract: We present a mathematical construction for the restricted Boltzmann machine (RBM) that does not require specifying the number of hidden units. In fact, the hidden layer size is adaptive and can grow during training. This is obtained by first extending the RBM to be sensitive to the ordering of its hidden units. Then, with a carefully chosen definition of the energy function, we show that the limit of infinitely many hidden units is well defined. As with RBM, approximate maximum likelihood training can be perfo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
63
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 52 publications
(63 citation statements)
references
References 10 publications
0
63
0
Order By: Relevance
“…It can also be further approximated with the function ln(1 + e s ). This method has been used in the past to develop models such as the Infinite RBM and Rate-coded RBM in generative machine learning [53,54].…”
Section: Generative Bi-partite Modelmentioning
confidence: 99%
“…It can also be further approximated with the function ln(1 + e s ). This method has been used in the past to develop models such as the Infinite RBM and Rate-coded RBM in generative machine learning [53,54].…”
Section: Generative Bi-partite Modelmentioning
confidence: 99%
“…Max-norm regularization [16] was also used to suppress very large weights, the bounds for each i W  and i U  were 10 and 5 respectively. Côté and Larochelle [14] claims that results of learning are robust to the value of the hidden unit penalty i  . We have tried several different i  and find that smaller i  enables the model to grow to proper size faster at the beginning of learning.…”
Section: Evaluation Of the Modelsmentioning
confidence: 99%
“…Nair,et al [13] conceptually tie the weights of an infinite number of binary hidden units, and connect these sigmoid units with noisy rectified linear units (ReLUs) for better feature learning. More recently, Côté and Larochelle [14] have proposed a non-parametric model called the iRBM. By making the effective number of hidden units participating in the energy function change freely during training, the iRBM can automatically adjust the effective number of hidden units according to the data.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations