2019
DOI: 10.48550/arxiv.1909.11274
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Compression based bound for non-compressed network: unified generalization error analysis of large compressible deep neural network

Abstract: One of the biggest issues in deep learning theory is the generalization ability of networks with huge model size. The classical learning theory suggests that overparameterized models cause overfitting. However, practically used large deep models avoid overfitting, which is not well explained by the classical approaches. To resolve this issue, several attempts have been made. Among them, the compression based bound is one of the promising approaches. However, the compression based bound can be applied only to a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 36 publications
(47 reference statements)
0
3
0
Order By: Relevance
“…To close with a few pointers to the literature, as Lemma 3.2 is essentially a pruning bound, it is potentially of independent interest; see for instance the literature on lottery tickets and pruning (Frankle and Carbin, 2019;Frankle et al, 2020;Su et al, 2020). Secondly, there is already one generalization bound in the literature which exhibits spectral norms, due to Suzuki et al (2019); unfortunately, it also has an explicit dependence on network width.…”
Section: Direct Uniform Convergence Approach In Theorem 14mentioning
confidence: 99%
See 2 more Smart Citations
“…To close with a few pointers to the literature, as Lemma 3.2 is essentially a pruning bound, it is potentially of independent interest; see for instance the literature on lottery tickets and pruning (Frankle and Carbin, 2019;Frankle et al, 2020;Su et al, 2020). Secondly, there is already one generalization bound in the literature which exhibits spectral norms, due to Suzuki et al (2019); unfortunately, it also has an explicit dependence on network width.…”
Section: Direct Uniform Convergence Approach In Theorem 14mentioning
confidence: 99%
“…1. We can directly apply uniform convergence; for instance, this approach was followed by Suzuki et al (2019), and a more direct approach is followed here to prove Theorem 1.4. Unfortunately, it is unclear how this technique can avoid paying significantly for the high complexity of f .…”
Section: An Abstract Bound Via Data Augmentationmentioning
confidence: 99%
See 1 more Smart Citation