2021
DOI: 10.1088/1742-5468/abfa1e
|View full text |Cite
|
Sign up to set email alerts
|

The loss surfaces of neural networks with general activation functions

Abstract: The loss surfaces of deep neural networks have been the subject of several studies, theoretical and experimental, over the last few years. One strand of work considers the complexity, in the sense of local optima, of high dimensional random functions with the aim of informing how local optimisation methods may perform in such complicated settings. Prior work of Choromanska et al (2015) established a direct link between the training loss surfaces of deep multi-layer perceptron networks and spherical multi-spin … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 12 publications
(16 citation statements)
references
References 142 publications
(344 reference statements)
0
16
0
Order By: Relevance
“…However, the Hessian index constraints are formulated in the natural Hessian matrix ( 16 ), but our spectral calculations proceed from the rewritten form ( 21 ). We find however that we can indeed proceed much as in [ 13 ]. Recall the key Hessian matrix given in ( 16 ) by where , , G is Ginibre, and all are independent.…”
Section: The Asymptotic Complexitymentioning
confidence: 52%
See 4 more Smart Citations
“…However, the Hessian index constraints are formulated in the natural Hessian matrix ( 16 ), but our spectral calculations proceed from the rewritten form ( 21 ). We find however that we can indeed proceed much as in [ 13 ]. Recall the key Hessian matrix given in ( 16 ) by where , , G is Ginibre, and all are independent.…”
Section: The Asymptotic Complexitymentioning
confidence: 52%
“…We use multi-spin glasses in high dimensions as a toy model for neural network loss surfaces without any further justification, beyond that found in [ 1 , 13 ]. GANs are composed of two networks: generator ( G ) and discriminator ( D ).…”
Section: An Interacting Spin Glass Modelmentioning
confidence: 99%
See 3 more Smart Citations