2019
DOI: 10.48550/arxiv.1911.02151
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Information-Theoretic Generalization Bounds for SGLD via Data-Dependent Estimates

Abstract: In this work, we improve upon the stepwise analysis of noisy iterative learning algorithms initiated by Pensia, Jog, and Loh (2018) and recently extended by Bu, Zou, and Veeravalli (2019). Our main contributions are significantly improved mutual information bounds for Stochastic Gradient Langevin Dynamics via datadependent estimates. Our approach is based on the variational characterization of mutual information and the use of data-dependent priors that forecast the minibatch gradient based on a subset of the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
13
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(13 citation statements)
references
References 19 publications
0
13
0
Order By: Relevance
“…Recently, researchers [37,30,28] propose to bound the generalization error by the mutual information between output hypothesis and input samples. Negrea et al [25] further tighten the bound by designing a data-dependent prior. Following [25], Haghifam et al [14] obtain comparable results through conditional mutual information [31].…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Recently, researchers [37,30,28] propose to bound the generalization error by the mutual information between output hypothesis and input samples. Negrea et al [25] further tighten the bound by designing a data-dependent prior. Following [25], Haghifam et al [14] obtain comparable results through conditional mutual information [31].…”
Section: Related Workmentioning
confidence: 99%
“…Negrea et al [25] further tighten the bound by designing a data-dependent prior. Following [25], Haghifam et al [14] obtain comparable results through conditional mutual information [31]. Other related work [2,16,5,13] tightens the information theoretical generalization bounds from different perspectives.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations