2021 IEEE International Symposium on Information Theory (ISIT) 2021
DOI: 10.1109/isit45174.2021.9518043
|View full text |Cite
|
Sign up to set email alerts
|

Information-Theoretic Bounds on the Moments of the Generalization Error of Learning Algorithms

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
4

Relationship

2
6

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 12 publications
0
5
0
Order By: Relevance
“…PAC-Bayesian generalization error bounds: First proposed by [6], [51], [52], PAC-Bayesian analysis provides high probability bounds on the generalization error in terms of KL divergence between the data-dependent posterior induced by the learning algorithm and a data-free prior that can be chosen arbitrarily [53]. There are multiple ways to generalize the standard PAC-Bayesian bounds, including using different information measures other than the KL divergence [54]- [58] and considering data-dependent priors (prior depends on the training data) [13], [59]- [63] or distribution-dependent priors (prior depends on data-generating distribution) [64]- [67]. In [68], a more general PAC-Bayesian framework is proposed, which provides a high probability bound on the convex function of the expected population and empirical risk with respect to the posterior distribution, whereas in [69] the connection between Bayesian inference and PAC-Bayesian theorem is explored by considering Gibbs posterior and negative log loss function.…”
Section: F Other Related Workmentioning
confidence: 99%
“…PAC-Bayesian generalization error bounds: First proposed by [6], [51], [52], PAC-Bayesian analysis provides high probability bounds on the generalization error in terms of KL divergence between the data-dependent posterior induced by the learning algorithm and a data-free prior that can be chosen arbitrarily [53]. There are multiple ways to generalize the standard PAC-Bayesian bounds, including using different information measures other than the KL divergence [54]- [58] and considering data-dependent priors (prior depends on the training data) [13], [59]- [63] or distribution-dependent priors (prior depends on data-generating distribution) [64]- [67]. In [68], a more general PAC-Bayesian framework is proposed, which provides a high probability bound on the convex function of the expected population and empirical risk with respect to the posterior distribution, whereas in [69] the connection between Bayesian inference and PAC-Bayesian theorem is explored by considering Gibbs posterior and negative log loss function.…”
Section: F Other Related Workmentioning
confidence: 99%
“…Robust PLS: Robustness of PLS is a widely discussed issue in the self-training literature. [Aminian et al, 2022] propose information-theoretic PLS robust towards covariate shift. [Lienen and Hüllermeier, 2021] label instances in the form of sets of probability distributions (credal sets), weakening the reliance on a single distribution.…”
Section: Related Workmentioning
confidence: 99%
“…Information-theoretic generalization error bounds using other information quantities are also studied, such as α-Rényi divergence and maximal leakage (Esposito et al, 2021), Jensen-Shannon divergence (Aminian et al, 2021b), power divergence (Aminian et al, 2021c), and Wasserstein distance (Lopez and Jog, 2018;Wang et al, 2019). An exact characterization of the generalization error for the Gibbs algorithm is provided in (Aminian et al, 2021a).…”
Section: Related Workmentioning
confidence: 99%