2020 IEEE Information Theory Workshop (ITW) 2021
DOI: 10.1109/itw46852.2021.9457642
|View full text |Cite
|
Sign up to set email alerts
|

Jensen-Shannon Information Based Characterization of the Generalization Error of Learning Algorithms

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

3
5

Authors

Journals

citations
Cited by 13 publications
(10 citation statements)
references
References 13 publications
0
10
0
Order By: Relevance
“…Information-theoretic generalization error bounds using other information quantities are also studied, such as α-Rényi divergence and maximal leakage (Esposito et al, 2021), Jensen-Shannon divergence (Aminian et al, 2021b), power divergence (Aminian et al, 2021c), and Wasserstein distance (Lopez and Jog, 2018;Wang et al, 2019). An exact characterization of the generalization error for the Gibbs algorithm is provided in (Aminian et al, 2021a). Using rate-distortion theory, Masiha et al (2021) and Bu et al (2020a) provide informationtheoretic generalization error upper bounds for model misspecification and model compression.…”
Section: Related Workmentioning
confidence: 99%
“…Information-theoretic generalization error bounds using other information quantities are also studied, such as α-Rényi divergence and maximal leakage (Esposito et al, 2021), Jensen-Shannon divergence (Aminian et al, 2021b), power divergence (Aminian et al, 2021c), and Wasserstein distance (Lopez and Jog, 2018;Wang et al, 2019). An exact characterization of the generalization error for the Gibbs algorithm is provided in (Aminian et al, 2021a). Using rate-distortion theory, Masiha et al (2021) and Bu et al (2020a) provide informationtheoretic generalization error upper bounds for model misspecification and model compression.…”
Section: Related Workmentioning
confidence: 99%
“…We can also apply the average joint distribution approach to the Jensen-Shannon divergence based upper bound in [11].…”
Section: Jensen-shannon Divergence Based Upper Boundmentioning
confidence: 99%
“…Bu et al [8] have derived tighter generalization error bounds based on individual sample mutual information. The generalization error bounds based on other information measures such as α-Réyni divergence [9], maximal leakage [10], Jensen-Shannon divergence [11], Wasserstein distances [12,13] and individual sample Wasserstein distance [14] are also considered. Chaining mutual information technique is proposed in [15] and [16] to further improve the mutual information-based bound.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Bounds using chaining mutual information have been proposed in [7]. Other authors have also constructed information-theoretic based average generalization error bounds using quantities such as α-Réyni divergence, f -divergence, Jensen-Shannon divergences, Wasserstein distances, or maximal leakage (see [8], [9], [10], [11], or [12]).…”
Section: Introductionmentioning
confidence: 99%