2022
DOI: 10.48550/arxiv.2203.14661
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Random matrix analysis of deep neural network weight matrices

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 73 publications
0
4
0
Order By: Relevance
“…Note that the idea of implicit regularisation of neural networks via stochastic gradient descent pre-dates this work by several years [NTS14; Ney+17a; Ney+17b; Ney17]. Finally, we mention [TSR22] in which the spectra of random and trained neural network weight matrices was analysed but on the local scale, rather than the global scale pursued by [MM18]. This work followed on from our own in Chapter 7 [BGK22] and similarly discovered the robust presence of universal GOE random matrix spacing statistics in the spectra.…”
Section: Spectra Of Neural Networkmentioning
confidence: 98%
See 1 more Smart Citation
“…Note that the idea of implicit regularisation of neural networks via stochastic gradient descent pre-dates this work by several years [NTS14; Ney+17a; Ney+17b; Ney17]. Finally, we mention [TSR22] in which the spectra of random and trained neural network weight matrices was analysed but on the local scale, rather than the global scale pursued by [MM18]. This work followed on from our own in Chapter 7 [BGK22] and similarly discovered the robust presence of universal GOE random matrix spacing statistics in the spectra.…”
Section: Spectra Of Neural Networkmentioning
confidence: 98%
“…At the macroscopic scale, there are results relevant to neural networks, for example [PSG18; Pas20] consider random neural networks with Gaussian weights and establish results that are generalised to arbitrary distributions with optimal conditions, so demonstrating universality. On the microscopic scale, our work in Chapter 7 provided the first evidence of universal random matrix theory statistics in neural networks and was subsequently to the weight matrices of neural networks in [TSR22], but no prior work has considered the implications of these statistics, that being the central contribution of Chapter 8. Our main mathematical result is a significant generalisation of the Hessian spectral outlier result recently presented by [GZR20].…”
Section: Discovery Of Rmt Universality In Loss Surfaces and Consequen...mentioning
confidence: 99%
“…On the microscopic scale, [BGK22] provided the first experimental demonstration of the presence of universal local random matrix statistics deep neural networks, specifically in the Hessians and Gauss-Newton matrices of their loss surfaces. This work has recently been extended to the weight matrices of neural networks [TSR22]. This paper explores the consequences of random matrix universality in deep neural networks.…”
Section: Introductionmentioning
confidence: 99%
“…On the microscopic scale, [BGK22] provided the first experimental demonstration of the presence of universal local random matrix statistics DNNs, specifically in the Hessians and Gauss-Newton matrices of their loss surfaces. This work has recently been extended to the weight matrices of neural networks [TSR22].…”
Section: Introductionmentioning
confidence: 99%