2021
DOI: 10.48550/arxiv.2109.05642
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the Impact of Spurious Correlation for Out-of-distribution Detection

Abstract: Modern neural networks can assign high confidence to inputs drawn from outside the training distribution, posing threats to models in real-world deployments. While much research attention has been placed on designing new out-of-distribution (OOD) detection methods, the precise definition of OOD is often left in vagueness and falls short of the desired notion of OOD in reality. In this paper, we present a new formalization and model the data shifts by taking into account both the invariant and environmental (sp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 21 publications
0
2
0
Order By: Relevance
“…Attributes such as gender in the mentioned example are called protected variables. In OOD detection literature, a recent work [148] systematically investigates how spurious correlation in the training set impacts OOD detection. The results suggest that the OOD detection performance is severely worsened when the correlation between spurious features and labels is increased in the training set.…”
Section: Fairness and Biases Of Modelsmentioning
confidence: 99%
“…Attributes such as gender in the mentioned example are called protected variables. In OOD detection literature, a recent work [148] systematically investigates how spurious correlation in the training set impacts OOD detection. The results suggest that the OOD detection performance is severely worsened when the correlation between spurious features and labels is increased in the training set.…”
Section: Fairness and Biases Of Modelsmentioning
confidence: 99%
“…We also provide a more fine-grained categorization of distributions for the purpose of thoroughly evaluating an algorithm. Specifically, we divide distributions into four groups: training ID, covariate-shifted ID, near-OOD, and far-OOD (the latter two are inspired by a recent study [10]). Figure 1-a shows example images from the DIGITS benchmark: the covariate-shifted images contain the same semantics as the training images, i.e., digits from 0 to 9, and should be classified as ID, whereas the two OOD groups clearly differ in semantics but represent two different levels of covariate shift.…”
Section: Introductionmentioning
confidence: 99%