2021
DOI: 10.48550/arxiv.2102.02016
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Information-Theoretic Bounds on the Moments of the Generalization Error of Learning Algorithms

Abstract: Generalization error bounds are critical to understanding the performance of machine learning models. In this work, building upon a new bound of the expected value of an arbitrary function of the population and empirical risk of a learning algorithm, we offer a more refined analysis of the generalization behaviour of a machine learning models based on a characterization of (bounds) to their generalization error moments. We discuss how the proposed bounds -which also encompass new bounds to the expected general… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 11 publications
(22 reference statements)
0
1
0
Order By: Relevance
“…[28] provides tighter bounds by considering the individual sample mutual information, [25,29] propose using chaining mutual information, and [30,31,32] advocate the conditioning and processing techniques. Information-theoretic generalization error bounds using other information quantities are also studied, such as, f -divergence [33], α-Rényi divergence and maximal leakage [34,35], and Jensen-Shannon divergence [36,37]. Using rate-distortion theory, [38,39,40] provide information-theoretic generalization error upper bounds for model misspecification and model compression.…”
Section: Related Workmentioning
confidence: 99%
“…[28] provides tighter bounds by considering the individual sample mutual information, [25,29] propose using chaining mutual information, and [30,31,32] advocate the conditioning and processing techniques. Information-theoretic generalization error bounds using other information quantities are also studied, such as, f -divergence [33], α-Rényi divergence and maximal leakage [34,35], and Jensen-Shannon divergence [36,37]. Using rate-distortion theory, [38,39,40] provide information-theoretic generalization error upper bounds for model misspecification and model compression.…”
Section: Related Workmentioning
confidence: 99%