2022 IEEE International Symposium on Information Theory (ISIT) 2022
DOI: 10.1109/isit50566.2022.9834474
|View full text |Cite
|
Sign up to set email alerts
|

Tighter Expected Generalization Error Bounds via Convexity of Information Measures

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 17 publications
0
5
0
Order By: Relevance
“…Similar bounds on the EGE were obtained in (Bu, Zou, and Veeravalli 2020;Chu and Raginsky 2023;Hafez-Kolahi et al 2020;Hellström and Durisi 2020) and references therein. Other information measures such as the Wasserstein distance (Aminian et al 2022;Lopez and Jog 2018;Wang et al 2019), maximal leakage (Esposito, Gastpar, and Issa 2020;Issa, Esposito, and Gastpar 2019), mutual f -information (Masiha, Gohari, and Yassaee 2023), and Jensen-Shannon divergence were used for providing upper bounds on EGE as well. In (Duchi, Glynn, and Namkoong 2021), the notion of closeness of probability measures with respect to a reference measure in terms of statistical distances was used.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Similar bounds on the EGE were obtained in (Bu, Zou, and Veeravalli 2020;Chu and Raginsky 2023;Hafez-Kolahi et al 2020;Hellström and Durisi 2020) and references therein. Other information measures such as the Wasserstein distance (Aminian et al 2022;Lopez and Jog 2018;Wang et al 2019), maximal leakage (Esposito, Gastpar, and Issa 2020;Issa, Esposito, and Gastpar 2019), mutual f -information (Masiha, Gohari, and Yassaee 2023), and Jensen-Shannon divergence were used for providing upper bounds on EGE as well. In (Duchi, Glynn, and Namkoong 2021), the notion of closeness of probability measures with respect to a reference measure in terms of statistical distances was used.…”
Section: Related Workmentioning
confidence: 99%
“…The expected generalization error (GE) is a central workhorse for the analysis of generalization capabilities of machine learning algorithms, see for instance (Aminian et al , 2022Chu and Raginsky 2023;Xu and Raginsky 2017) and (Perlaza et al 2023). In a nutshell, the GE characterizes the ability of the learning algorithm to correctly find patterns in datasets that are not available during the training stage.…”
Section: Introductionmentioning
confidence: 99%
“…[27] provides tighter bounds by considering the individual sample mutual information, [28], [29] propose using chaining mutual information, and [30]- [32] advocate the conditioning and processing techniques. Information-theoretic generalization error bounds using other information quantities are also studied, such as f -divergence [33], α-Rényi divergence and maximal leakage [34], [35], Jensen-Shannon divergence [36], [37] and Wasserstein distance [38]- [41]. In [42], upper bounds in terms of mutual information are obtained by employing coupling and chaining techniques in the space of probability measures.…”
Section: F Other Related Workmentioning
confidence: 99%
“…One of the traditional performance metrics to evaluate the generalization capabilities of the Gibbs algorithm is the generalization error. When the reference measure is a probability measure, a closed-form expression for the generalization error of the Gibbs algorithm is presented in [9], while upper bounds have been derived in [16], [21], [28]- [34], [39]- [52], and references therein. In this work, a new performance metric coined sensitivity, which quantifies the variations of the expected empirical risk due to deviations from the solution of the ERM-RER problem is introduced.…”
Section: Introductionmentioning
confidence: 99%