2020
DOI: 10.48550/arxiv.2005.01856
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Selecting Data Augmentation for Simulating Interventions

Abstract: Machine learning models trained with purely observational data and the principle of empirical risk minimization (Vapnik, 1992) can fail to generalize to unseen domains. In this paper, we focus on the case where the problem arises through spurious correlation between the observed domains and the actual task labels. We find that many domain generalization methods do not explicitly take this spurious correlation into account. Instead, especially in more application-oriented research areas like medical imaging or … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 21 publications
0
7
0
Order By: Relevance
“…This assumption is not feasible if we do not have access to this model. Similar connections of causality and data augmentation have been made in (Ilse et al, 2020). It is also possible to train on interventional images x IV , i.e., generating a single image per sampled noise vector.…”
Section: Training An Invariant Classifiermentioning
confidence: 90%
“…This assumption is not feasible if we do not have access to this model. Similar connections of causality and data augmentation have been made in (Ilse et al, 2020). It is also possible to train on interventional images x IV , i.e., generating a single image per sampled noise vector.…”
Section: Training An Invariant Classifiermentioning
confidence: 90%
“…These methods essentially minimize the differences among domain specific risks. Moreover, data augmentation can also improve the generalizability of DNNs from the data perspective [12,26]. However, it requires prior knowledge on the differences between training and test domains, which are not be available in OOD prediction.…”
Section: An Effective Fix For Irmmentioning
confidence: 99%
“…. We measure the policy distance using Jensen-Shannon Divergence because it is a lower bound for the joint empirical risk across nonaugmented and augmented observations (Ilse et al, 2020;Raileanu et al, 2020). The policy distances are remarkably reduced when using augmentation in spite of delayed usage at (10, 25), (20,25) We analyze about two samples, which are trained on Bigfish environments, for verifying when generalization occurs.…”
Section: Augmentation During Rl Trainingmentioning
confidence: 99%