1997
DOI: 10.1051/jp1:1997147
|View full text |Cite
|
Sign up to set email alerts
|

Replica Symmetry Breaking and the Kuhn-Tucker Cavity Method in Simple and Multilayer Perceptrons

Abstract: Within a Kuhn-Tucker cavity method introduced in a former paper, we study optimal stability learning for situations, where in the replica formalism the replica symmetry may be broken, namely (i) the case of a simple perceptron above the critical loading, and (ii) the case of two-layer AND-perceptrons, if one learns with maximal stability. We find that the deviation of our cavity solution from the replica symmetric one in these cases is a clear indication of the necessity of replica symmetry breaking. In any ca… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2000
2000
2006
2006

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 25 publications
0
7
0
Order By: Relevance
“…It is however worth to examine it. The corresponding value of θ as a function of ε and κ follows form (18), and the relation between α, θ, κ and c from (17). We find:…”
Section: A Replica Calculationmentioning
confidence: 84%
See 3 more Smart Citations
“…It is however worth to examine it. The corresponding value of θ as a function of ε and κ follows form (18), and the relation between α, θ, κ and c from (17). We find:…”
Section: A Replica Calculationmentioning
confidence: 84%
“…In order to circumvent the RS approximation, we determine the training error E t (α, ε) using the Kuhn-Tucker (KT) cavity method proposed by Gerl and Krey [18], that we generalize here to the case of a perceptron with a threshold learning a training set with a biased probability of targets given by (5). Contrary to the RS solution, this cavity method has been shown to overestimate the training error [18]. Consequently, the results allow us to deduce an upper bound for the number of perceptrons needed by the tilinglike procedure to converge.…”
Section: B Kuhn-tucker Cavity Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…It gives the lowest bound to E t . In the limit of large training set size α, we obtain: In order to obtain a lower bound to α arch c (k) we used the Kuhn-Tucker cavity method [14,11,12], which gives an upper bound to E t . As a result of both calculations, we can bound α alg c (k, GD):…”
Section: Gardner-derrida Cost Functionmentioning
confidence: 99%