2021
DOI: 10.48550/arxiv.2106.04110
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A self consistent theory of Gaussian Processes captures feature learning effects in finite CNNs

Gadi Naveh,
Zohar Ringel

Abstract: Deep neural networks (DNNs) in the infinite width/channel limit have received much attention recently, as they provide a clear analytical window to deep learning via mappings to Gaussian Processes (GPs). Despite its theoretical appeal, this viewpoint lacks a crucial ingredient of deep learning in finite DNNs, laying at the heart of their success -feature learning. Here we consider DNNs trained with noisy gradient descent on a large training set and derive a self consistent Gaussian Process theory accounting fo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
12
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(13 citation statements)
references
References 32 publications
1
12
0
Order By: Relevance
“…The latter captured several strong finite DNN effects. Some empirical result on real world CNN verified the assumptions underlying this approach, namely, that the trained CNN outputs fluctuate in a nearly Gaussian manner [27].…”
Section: Introductionmentioning
confidence: 75%
See 4 more Smart Citations
“…The latter captured several strong finite DNN effects. Some empirical result on real world CNN verified the assumptions underlying this approach, namely, that the trained CNN outputs fluctuate in a nearly Gaussian manner [27].…”
Section: Introductionmentioning
confidence: 75%
“…We refer to this as kernel stability. We expect it to hold more generally, as it is essentially the statement that a latent kernel, which encompasses the feature learning effects [1,27], is stable to changing a few data points at large n.…”
Section: A Single Activated Layermentioning
confidence: 99%
See 3 more Smart Citations