2020
DOI: 10.48550/arxiv.2003.00688
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Out-of-Distribution Generalization via Risk Extrapolation (REx)

David Krueger,
Ethan Caballero,
Joern-Henrik Jacobsen
et al.

Abstract: Generalizing outside of the training distribution is an open challenge for current machine learning systems. A weak form of out-of-distribution (OoD) generalization is the ability to successfully interpolate between multiple observed distributions. One way to achieve this is through robust optimization, which seeks to minimize the worstcase risk over convex combinations of the training distributions. However, a much stronger form of OoD generalization is the ability of models to extrapolate beyond the distribu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

5
118
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 59 publications
(123 citation statements)
references
References 22 publications
(42 reference statements)
5
118
0
Order By: Relevance
“…A plethora of algorithms are proposed: learning invariant representation across domains [7,21,38,20], minimizing the weighted combination of risks from training domains [35], using different risk penalty terms to facilitate invariance prediction [1,17], causal inference approaches [31], and forcing the learned representation different from a set of pre-defined biased representations [2], mixup-based approaches [48,41,26], etc. A recent study [10] shows that no domain generalization methods achieve superior performance than ERM across a broad range of datasets.…”
Section: Discussion and Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…A plethora of algorithms are proposed: learning invariant representation across domains [7,21,38,20], minimizing the weighted combination of risks from training domains [35], using different risk penalty terms to facilitate invariance prediction [1,17], causal inference approaches [31], and forcing the learned representation different from a set of pre-defined biased representations [2], mixup-based approaches [48,41,26], etc. A recent study [10] shows that no domain generalization methods achieve superior performance than ERM across a broad range of datasets.…”
Section: Discussion and Related Workmentioning
confidence: 99%
“…Extension: Empirical Validation of Theoretical Analysis. To further validate our analysis above, we comprehensively evaluate the OOD detection performance of models that are trained with recent prominent domain invariance learning objectives [1,2,17,7,21,35] (Section E in Appendix). The results align with our theoretical analysis.…”
Section: Ood Typementioning
confidence: 99%
See 2 more Smart Citations
“…In an ideal case, even simple k-NN based method can perform well. However, when the variation among class-conditionals of the same class is large, i.e., the closest conditional distribution to T (X|Y = y) is some S(X|Y = y ) of class y = y (Figure 1 Right), aforementioned methods may not perform well (Krueger et al, 2020).…”
Section: Introductionmentioning
confidence: 99%