2018
DOI: 10.1007/s10994-018-5696-2
|View full text |Cite
|
Sign up to set email alerts
|

On better training the infinite restricted Boltzmann machines

Abstract: The infinite restricted Boltzmann machine (iRBM) is an extension of the classic RBM. It enjoys a good property of automatically deciding the size of the hidden layer according to specific training data. With sufficient training, the iRBM can achieve a competitive performance with that of the classic RBM. However, the convergence of learning the iRBM is slow, due to the fact that the iRBM is sensitive to the ordering of its hidden units, the learned filters change slowly from the left-most hidden unit to right.… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7
2

Relationship

2
7

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 15 publications
0
7
0
Order By: Relevance
“…The second experiment investigated the robustness of the proposed model when the angular sampling rates at test phase were different from that at training phase. To accelerate the learning of Dis-iRBMs, we used a new training strategy referred as “RP” training [ 34 ] to train the Dis-iRBMs in all experiments.…”
Section: Resultsmentioning
confidence: 99%
“…The second experiment investigated the robustness of the proposed model when the angular sampling rates at test phase were different from that at training phase. To accelerate the learning of Dis-iRBMs, we used a new training strategy referred as “RP” training [ 34 ] to train the Dis-iRBMs in all experiments.…”
Section: Resultsmentioning
confidence: 99%
“…The second procedure is supervised fine-tuning of the whole networks by back propagation algorithm. In this procedure, all of the hidden layers are considered as a whole and model parameters are adjusted to decrease the training error [25].…”
Section: Deep Belief Networkwith Gaussian-bernoulli Rbmmentioning
confidence: 99%
“…The ADAGRAD [43], a per‐dimensional learning rate was also used as [32] does. To accelerate the learning speed and achieve better generalisation, a new training strategy referred as ‘RP training’ [44] was adopted here for all the experiments.…”
Section: Experiments On Radar Hrrp Recognitionmentioning
confidence: 99%