2019
DOI: 10.1088/1751-8121/ab3f3f
|View full text |Cite
|
Sign up to set email alerts
|

Minimal model of permutation symmetry in unsupervised learning

Abstract: Permutation of any two hidden units yields invariant properties in typical deep generative neural networks. This permutation symmetry plays an important role in understanding the computation performance of a broad class of neural networks with two or more hidden units. However, a theoretical study of the permutation symmetry is still lacking. Here, we propose a minimal model with only two hidden units in a restricted Boltzmann machine, which aims to address how the permutation symmetry affects the critical lea… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
11
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1
1

Relationship

2
5

Authors

Journals

citations
Cited by 14 publications
(11 citation statements)
references
References 30 publications
0
11
0
Order By: Relevance
“…Now we calculate the free energy of the model for the extensive-load case α = P/N ∼ O(1). To derive a typical behavior of the model, we need to perform a disorder average of ln Z, which can be tackled by the replica method: −βf = lim n→0,N →∞ ln Z n nN (e.g., see [24,25]). In essence, n copies of the original system are introduced.…”
Section: Arxiv:210314317v1 [Cond-matdis-nn] 26 Mar 2021mentioning
confidence: 99%
“…Now we calculate the free energy of the model for the extensive-load case α = P/N ∼ O(1). To derive a typical behavior of the model, we need to perform a disorder average of ln Z, which can be tackled by the replica method: −βf = lim n→0,N →∞ ln Z n nN (e.g., see [24,25]). In essence, n copies of the original system are introduced.…”
Section: Arxiv:210314317v1 [Cond-matdis-nn] 26 Mar 2021mentioning
confidence: 99%
“…By asking a minimal data size to trigger learning in a two-layer neural network, namely a restricted Boltzmann machine (RBM) [15], a recent study claimed that sensory inputs (or data streams) are able to drive a series of phase transitions related to broken inherent-symmetries of the model [16]. However, this model does not assume any prior knowledge during learning, therefore the impact of priors on the learning remains unexplained.…”
mentioning
confidence: 99%
“…From a neural network perspective, the interplay between the prior and the likelihood function of data can be captured by synaptic weights. These synaptic weights are modeled by feedforward connections in a RBM [15,16,18]. More precisely, the RBM is a two-layer neural network where there do not exist intra-layer connections.…”
mentioning
confidence: 99%
See 2 more Smart Citations