2022 IEEE Information Theory Workshop (ITW) 2022
DOI: 10.1109/itw54588.2022.9965761
|View full text |Cite
|
Sign up to set email alerts
|

Fast Rate Generalization Error Bounds: Variations on a Theme

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 10 publications
0
3
0
Order By: Relevance
“…To address this limitation, several works, inspired by some PAC-Bayesian literature, have proposed fast-rate information-theoretic bounds [24,25,64], demonstrating a good characterization in some instances of non-convex settings such as deep learning. More recently, [65,66,71] present IOMI bounds for the Gaussian mean estimation problem, that achieve optimal convergence rates, contrasting previous bounds [9,70].…”
Section: Introductionmentioning
confidence: 94%
See 1 more Smart Citation
“…To address this limitation, several works, inspired by some PAC-Bayesian literature, have proposed fast-rate information-theoretic bounds [24,25,64], demonstrating a good characterization in some instances of non-convex settings such as deep learning. More recently, [65,66,71] present IOMI bounds for the Gaussian mean estimation problem, that achieve optimal convergence rates, contrasting previous bounds [9,70].…”
Section: Introductionmentioning
confidence: 94%
“…Recently, [65,66] use the unexpected excess risk as the DV auxiliary function and invoke the (η, c)central condition to establish some optimal-rate bounds for specific learning problems, e.g., Gaussian mean estimation. This again highlights the significance of selecting appropriate DV auxiliary functions and corresponding assumptions tailored to different learning problems.…”
Section: Extensionsmentioning
confidence: 99%
“…Subsequent to our work, Wu et al [2022] prove fast-rate conditional generalization bounds based on the (𝑣, 𝜂)-central assumption, which is weaker than the Bernstein condition and captures unbounded and subgaussian losses.…”
Section: Compression Scheme Priorsmentioning
confidence: 94%