2022
DOI: 10.1016/j.physa.2021.126742
|View full text |Cite
|
Sign up to set email alerts
|

Appearance of Random Matrix Theory in deep learning

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 26 publications
0
8
0
Order By: Relevance
“…As we have discussed at length hitherto, we conjecture that a local law is reasonable assumption to make on random matrices arising in deep neural networks. In particular in Chapter 7 [BGK22] we demonstrated universal local random matrix theory statistics not just for Hessians of deep networks but also for Generalised Gauss-Newton matrices. Our aim here is to demonstrate how a local law on Ĥt dramatically simplifies the statistics of (8.136).…”
Section: Implications For Curvature From Local Lawsmentioning
confidence: 94%
See 2 more Smart Citations
“…As we have discussed at length hitherto, we conjecture that a local law is reasonable assumption to make on random matrices arising in deep neural networks. In particular in Chapter 7 [BGK22] we demonstrated universal local random matrix theory statistics not just for Hessians of deep networks but also for Generalised Gauss-Newton matrices. Our aim here is to demonstrate how a local law on Ĥt dramatically simplifies the statistics of (8.136).…”
Section: Implications For Curvature From Local Lawsmentioning
confidence: 94%
“…The theoretical picture that has emerged is that, for very general random matrices, when universal local eigenvalue statistics are observed in random matrices, it is due to the mechanism of short time scale relaxation of local statistics under Dyson Brownian Motion made possible by a local law. In Chapter 7 [BGK22] we observed that universal local eigenvalue statistics do indeed appear to be present in the Hessian of real, albeit quite small, deep neural networks. Given all of this context, we propose that a local law assumption of some kind is reasonable for deep neural network Hessians and not particularly restrictive.…”
Section: Justification and Motivation Of Quementioning
confidence: 95%
See 1 more Smart Citation
“…Challenging the above-mentioned works, an experimental line of work has demonstrated convincingly that special RMT ensembles like the GOE do not appear to be present in DNNs [Pap18,Gra20,BGK22], for example as their Hessians. In addition, there have been challenges in the literature to the practical relevance of spin glass loss surface results for DNNs [BJSG+19].…”
Section: Introductionmentioning
confidence: 95%
“…where Z β is the normalization constant. This distribution has found many applications, to study eigenvalue statistics in spin systems [25,[28][29][30][31][32][33][34][35], in triangular billiards [36], in the Hessians of artificial neural networks [37], in Sachdev-Ye-Kitaev model [38][39][40][41][42], in quantum field theory [43], to quantify symmetries in various complex systems [44,45].…”
mentioning
confidence: 99%