2019
DOI: 10.1088/1751-8121/ab3709
|View full text |Cite
|
Sign up to set email alerts
|

Generalization from correlated sets of patterns in the perceptron

Abstract: Generalization is a central aspect of learning theory. Here, we propose a framework that explores an auxiliary task-dependent notion of generalization, and attempts to quantitatively answer the following question: given two sets of patterns with a given degree of dissimilarity, how easily will a network be able to "unify" their interpretation? This is quantified by the volume of the configurations of synaptic weights that classify the two sets in a similar manner. To show the applicability of our idea in a con… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3

Relationship

3
6

Authors

Journals

citations
Cited by 14 publications
(10 citation statements)
references
References 39 publications
(71 reference statements)
0
10
0
Order By: Relevance
“…For , the recurrence Equation ( 51 ) becomes By substituting given by Equation ( 53 ) and solving the recurrence we obtain, after some algebra, where is the second structure parameter that is defined in Equation ( 18 ). Finally, by combining the leading order expansions ( 50 ) and the moments ( 53 ) and ( 55 ), and by rescaling, as in Equation ( 31 ), we have the following explicit expressions for the two main metrics of separability as functions of the multiplicity k and the structure parameters and : For data that are structured as pairs of points, , Equation ( 56 ) gives the storage capacity of an ensemble of segments; this special result was first obtained, by means of replica calculations, in [ 44 ], and it was then rediscovered in other contexts in [ 8 , 45 ].…”
Section: Computation Of Compact Metrics Of Linear Separabilitymentioning
confidence: 99%
“…For , the recurrence Equation ( 51 ) becomes By substituting given by Equation ( 53 ) and solving the recurrence we obtain, after some algebra, where is the second structure parameter that is defined in Equation ( 18 ). Finally, by combining the leading order expansions ( 50 ) and the moments ( 53 ) and ( 55 ), and by rescaling, as in Equation ( 31 ), we have the following explicit expressions for the two main metrics of separability as functions of the multiplicity k and the structure parameters and : For data that are structured as pairs of points, , Equation ( 56 ) gives the storage capacity of an ensemble of segments; this special result was first obtained, by means of replica calculations, in [ 44 ], and it was then rediscovered in other contexts in [ 8 , 45 ].…”
Section: Computation Of Compact Metrics Of Linear Separabilitymentioning
confidence: 99%
“…2 and 3 report the numerical solutions of these saddle-point equations for the volumes, respectively, (2) and (3). The limiting values of α at the saddle point for q M → 1 can be worked out analytically (see [12,33]); they are, respectively,…”
Section: The Rs Solution and Its Stabilitymentioning
confidence: 99%
“…Integrating data structure within the framework of statistical mechanics is relatively straightforward and usually follows two steps: (i) define a generative model for the data, given in terms of a nonfactorized joint probability distribution P (X p , Y p ); (ii) compute averages over the measure P (the "disorder"); this is what was done for instance in [5,12,42]. How to best address data dependence in the SLT formalism, instead, is a debated issue.…”
Section: Constrained Models Of Structured Datamentioning
confidence: 99%