2021
DOI: 10.48550/arxiv.2110.10832
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Ensemble of Averages: Improving Model Selection and Boosting Performance in Domain Generalization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(8 citation statements)
references
References 25 publications
0
8
0
Order By: Relevance
“…the in-domain strategy by Gulrajani & Lopez-Paz (2020)) to select the best hyper-parameters, and report the average performance and standard deviation across 5 random seeds. Baselines : We compare our method against standard ERM training, which has proven to be a frustratingly difficult baseline (Gulrajani & Lopez-Paz, 2020), and also against several state of the art methods on this benchmark -SWAD (Cha et al, 2021), MIRO (Cha et al, 2022) and SMA (Arpit et al, 2021). Finally, we show that our approach can be effectively integrated with stochastic weight averaging to obtain further gains.…”
Section: Ood Generalization In a Real World Settingmentioning
confidence: 98%
See 1 more Smart Citation
“…the in-domain strategy by Gulrajani & Lopez-Paz (2020)) to select the best hyper-parameters, and report the average performance and standard deviation across 5 random seeds. Baselines : We compare our method against standard ERM training, which has proven to be a frustratingly difficult baseline (Gulrajani & Lopez-Paz, 2020), and also against several state of the art methods on this benchmark -SWAD (Cha et al, 2021), MIRO (Cha et al, 2022) and SMA (Arpit et al, 2021). Finally, we show that our approach can be effectively integrated with stochastic weight averaging to obtain further gains.…”
Section: Ood Generalization In a Real World Settingmentioning
confidence: 98%
“…Tackling the OOD robustness problem, Thomas et al (2021) and Matsuura & Harada (2020) first cluster training examples into "pseudo-domains", after which standard domain generalization techniques are used. Another recent line of works propose using model averaging (Cha et al, 2021;Li et al, 2022) and/or ensembling (Arpit et al, 2021) for better OOD generalization. These techniques are complementary to our contribution, and we demonstrate how they can benefit each other in our empirical evaluation.…”
Section: Domain Generalization and Ood Robustnessmentioning
confidence: 99%
“…PTMs for domain generalization. Methods leveraging pretraining models have shown promising improvements in domain generalization performance (Wiles et al, 2022;Arpit et al, 2021;Dong et al, 2022;Wortsman et al, 2022;Rame et al, 2022;Ramé et al, 2022). Among them, ensemble methods combined with PTMs show further advantages.…”
Section: Related Workmentioning
confidence: 99%
“…However, compared to the RHS of Equation 15, our proposed metrics could bring two advantages. Firstly, our proposed metrics are easier to approximate with finite samples in practice (as shown in Section 4.3 in the main paper and A.1 and A.2 in Appendix) while the estimation of KL divergence is challenging [Wang et al, 2021;Zhao et al, 2020]. Secondly, our proposed metrics have close connections with the error of models (as shown in Theorem 4.2 and Theorem 4.3), so that they are more befitting the evaluation of DG datasets for benchmarking DG algorithms.…”
Section: A3 Comparison Between the Proposed Metrics And Kullback-leib...mentioning
confidence: 99%
“…The distribution shift between training and test data may lead to the unreliable performance of most current approaches in practice. Hence, instead of generalization within the training distribution, the ability to generalize under distribution shift, namely domain generalization (DG) [Wang et al, 2021;, is of more critical significance in realistic scenarios. † Equal contribution * Corresponding Author ‡ The dataset can be found at https://www.dropbox.com/sh/u2bq2xo8sbax4pr/AADbhZJAy0AAbap76cg _ XkAfa?dl=0.…”
Section: Introductionmentioning
confidence: 99%