Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.272
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Model and Ignore Dataset Bias with Mixed Capacity Ensembles

Abstract: Many datasets have been shown to contain incidental correlations created by idiosyncrasies in the data collection process. For example, sentence entailment datasets can have spurious word-class correlations if nearly all contradiction sentences contain the word "not", and image recognition datasets can have telltale object-background correlations if dogs are always indoors. In this paper, we propose a method that can automatically detect and ignore these kinds of dataset-specific patterns, which we call datase… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
36
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 39 publications
(43 citation statements)
references
References 34 publications
0
36
0
Order By: Relevance
“…The results are satisfactory, especially when considering the simplicity and ef-ficiency of our approach. Moreover, the fact that a single configuration works well on 3 tasks is an indicator that our method has the potential to generalize on completely unknown OOD sets (Clark et al, 2020).…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…The results are satisfactory, especially when considering the simplicity and ef-ficiency of our approach. Moreover, the fact that a single configuration works well on 3 tasks is an indicator that our method has the potential to generalize on completely unknown OOD sets (Clark et al, 2020).…”
Section: Resultsmentioning
confidence: 99%
“…Following (Clark et al, 2019;Grand and Belinkov, 2019;Clark et al, 2020;Sanh et al, 2020), we tune our method hyperparameters on the OOD sets. As pointed out by (Clark et al, 2019(Clark et al, , 2020, this is not ideal since it assumes some prior knowledge of the OOD test sets. To best mitigate this impact, we follow the procedure of previous works and use the same hyper-parameters for all 3 tasks.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Belinkov et al (2019) used adversarial training to mitigate the hypothesis-only bias in textual entailment models. Clark et al (2020) adversarially trained a low and a high capacity model in an ensemble in order to ensure that the latter model is focusing on patterns that should generalize better. Dayanik and Padó (2020) Dai and Adel (2020) explored different entity substitution techniques for data augmentation tailored to NER.…”
Section: Mitigating Biasmentioning
confidence: 99%
“…To generalize to outof-distribution samples adaptively, the VQA model should own two capabilities: (1) overcoming negative language biases and (2) producing out-of-distribution answers by learning rules entailed in in-domain data. The prevailing OOD generalization methods [10,11,18,65] focus on enhancing the first capability, which achieves OOD generalization by explicitly mitigating the language biases. While the second capability, which directly endues VQA models the potentiality to generalize to out-of-distribution (i.e., unseen or rare) samples, has not been well explored.…”
mentioning
confidence: 99%