2020
DOI: 10.48550/arxiv.2003.13819
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Sharp Concentration Results for Heavy-Tailed Distributions

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…We will bound the right tail of i X i and the left tail is bounded similarly. Using the same argument as Lemma 3(a) from [BMP20], one has lim L→∞ c L = Var(X). Further, it follows from definition that c L is bounded in any bounded subset of (0, ∞) and that c L is a continuous function of Notice that from the condition that I(t) ≥ t γ /C for some C > 0, it follows that…”
Section: Sums Of Heavy Tailed Random Variablesmentioning
confidence: 92%
See 1 more Smart Citation
“…We will bound the right tail of i X i and the left tail is bounded similarly. Using the same argument as Lemma 3(a) from [BMP20], one has lim L→∞ c L = Var(X). Further, it follows from definition that c L is bounded in any bounded subset of (0, ∞) and that c L is a continuous function of Notice that from the condition that I(t) ≥ t γ /C for some C > 0, it follows that…”
Section: Sums Of Heavy Tailed Random Variablesmentioning
confidence: 92%
“…Next, we would like to analyze concentration for sums of independent instances of f ′ DE (η i ) and f ′′ DE (η i ). For that purpose, we use a general result on concentration of sums of independent random variables, which follows from [BMP20]. (The proof appears in Section 4.3) Lemma 4.5.…”
Section: Proofmentioning
confidence: 99%
“…The proof is given in Appendix C.2 and uses concentration results on heavy tailed distributions (Bakhshizadeh et al, 2020) to bound the operator norm of the matrix…”
Section: Effective Noise Levelmentioning
confidence: 99%
“…Heavy-tailed distributions are distributions whose moment-generating function diverges for all positive parameterizations ( [14]), and have been used as models of correlated systems ( [15,16,17,18,19,20]). In this work, we assume that the fully trained weight matrices of a fully connected neural network (FCNN) have power-law structure ( [11] and [1]), and using the compression framework of [2], prove bounds on generalization given this structure.…”
Section: Introductionmentioning
confidence: 99%
“…This results in a nonvacuous generalization bound, i.e., a bound that is less than one. The key idea of how to compress these matrices was inspired by the work of [16] and [15]. The work will be structured as follows:…”
Section: Introductionmentioning
confidence: 99%