2020
DOI: 10.1088/1751-8121/ab6a6f
|View full text |Cite
|
Sign up to set email alerts
|

Large deviation analysis of function sensitivity in random deep neural networks

Abstract: Mean field theory has been successfully used to analyze deep neural networks (DNN) in the infinite size limit. Given the finite size of realistic DNN, we utilize the large deviation theory and path integral analysis to study the deviation of functions represented by DNN from their typical mean field solutions. The parameter perturbations investigated include weight sparsification (dilution) and binarization, which are commonly used in model simplification, for both ReLU and sign activation functions. We find t… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
9
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 23 publications
(57 reference statements)
0
9
0
Order By: Relevance
“…However, it is also highly challenging due to the inherent recursiveness of computation and randomness in their architecture and/or computing elements. Existing theoretical studies of the function space of deep-layered machines are mostly based on the mean field approach, which allows for a sensitivity analysis of the functions realized by deep-layered machines due to input or parameter perturbations [4,[18][19][20].…”
mentioning
confidence: 99%
“…However, it is also highly challenging due to the inherent recursiveness of computation and randomness in their architecture and/or computing elements. Existing theoretical studies of the function space of deep-layered machines are mostly based on the mean field approach, which allows for a sensitivity analysis of the functions realized by deep-layered machines due to input or parameter perturbations [4,[18][19][20].…”
mentioning
confidence: 99%
“…Unlike Ref. [40,41], which studies perturbation around a ReLU network, our analysis aims to understand the critical properties of correlations. We consider correlations within the set of weights (w l i ) incoming to each neuron i, with all neurons identically distributed.…”
Section: Mean Field Analysis Of Signal Propagation With Correlated We...mentioning
confidence: 99%
“…A growing body of work has analyzed signal propagation in infinitely wide networks to understand forward-propagation in DNNs [35,36,37,38,39,40,41]. We mention a few results for ReLU networks.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Xiang et al [46] introduced the maximum sensitivity by a bounded disturbance on the nominal input to measure the maximum deviation of outputs. Li et al [47] studied the deviation of functions represented by DNN from their typical mean field solutions by the large deviation theory and path integral analysis, where the commonly used weight sparsification and binarization in model simplification were investigated under parameter perturbations. In [48], a provable Sensitivity-informed Provable Pruning (SiPPing) method of neural networks was suggested based on a ST-SA of measuring the importance of each weight for one layer.…”
Section: A the Stochastic Sensitivity Analysis Of Neural Networkmentioning
confidence: 99%