Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.74
|View full text |Cite
|
Sign up to set email alerts
|

Improving QA Generalization by Concurrent Modeling of Multiple Biases

Abstract: Existing NLP datasets contain various biases that models can easily exploit to achieve high performances on the corresponding evaluation sets. However, focusing on dataset-specific biases limits their ability to learn more generalizable knowledge about the task from more general data patterns. In this paper, we investigate the impact of debiasing methods for improving generalization and propose a general framework for improving the performance on both in-domain and out-of-domain datasets by concurrent modeling… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
10
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1

Relationship

3
4

Authors

Journals

citations
Cited by 8 publications
(10 citation statements)
references
References 36 publications
0
10
0
Order By: Relevance
“…The use of this information in the training objective improves the robustness of the model on adversarial datasets (He et al, 2019;Clark et al, 2019a;Utama et al, 2020a), i.e., datasets that contain counterexamples in which relying on the bias results in an incorrect prediction. In addition, it can also improve in-domain performances as well as generalization across various datasets that represent the same task (Wu et al, 2020a;Utama et al, 2020b).…”
Section: Artifacts In Nlp Datasetsmentioning
confidence: 99%
“…The use of this information in the training objective improves the robustness of the model on adversarial datasets (He et al, 2019;Clark et al, 2019a;Utama et al, 2020a), i.e., datasets that contain counterexamples in which relying on the bias results in an incorrect prediction. In addition, it can also improve in-domain performances as well as generalization across various datasets that represent the same task (Wu et al, 2020a;Utama et al, 2020b).…”
Section: Artifacts In Nlp Datasetsmentioning
confidence: 99%
“…The use of this information in the training objective improves the robustness of the model on adversarial datasets (He et al, 2019;Clark et al, 2019a;Utama et al, 2020a), i.e., datasets that contain counterexamples in which relying on the bias results in an incorrect prediction. In addition, it can also improve in-domain performances as well as generalization across various datasets that represent the same task (Wu et al, 2020a;Utama et al, 2020b).…”
Section: Artifacts In Nlp Datasetsmentioning
confidence: 99%
“…Therefore, while they improve the performance on the targeted adversarial sets, they may hurt the overall robustness. The recent work of Utama et al (2020b) and Wu et al (2020) are the exceptions in which they show that their proposed debiasing frameworks improve the overall robustness, and hence the generalization across different datasets in natural language understanding and question answering, respectively. Utama et al (2020b) propose a new framework that automatically recognizes biased training examples and does not require predefining bias types.…”
Section: Introductionmentioning
confidence: 99%
“…The majority of existing works improve the robustness against a given bias by proposing new methods or training paradigms (He et al, 2019;Clark et al, 2019;Mahabadi and Henderson, 2019;Utama et al, 2020a,b;Wu et al, 2020). The common component in such methods is a bias model that is trained to detect training examples that can be solved only using a bias.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation