2022
DOI: 10.48550/arxiv.2212.10563
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Debiasing NLP Models Without Demographic Information

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 0 publications
0
2
0
Order By: Relevance
“…Baselines. We consider popular baselines from prior work (Joshi et al, 2022;Orgad and Belinkov, 2022): weighting methods like DFL, DFL-nodemog, Product of Experts (Mahabadi et al, 2019;Orgad and Belinkov, 2022) and latent space removal methods like INLP (Ravfogel et al, 2020). We also include worst-group accuracy methods like GroupDRO, Subsampling (Sagawa et al, 2019(Sagawa et al, , 2020 from the machine learning literature, and a baseline RemoveToken that removes the treatment feature from input (see Supp C).…”
Section: Accuracy Of Feag Classifiersmentioning
confidence: 99%
See 1 more Smart Citation
“…Baselines. We consider popular baselines from prior work (Joshi et al, 2022;Orgad and Belinkov, 2022): weighting methods like DFL, DFL-nodemog, Product of Experts (Mahabadi et al, 2019;Orgad and Belinkov, 2022) and latent space removal methods like INLP (Ravfogel et al, 2020). We also include worst-group accuracy methods like GroupDRO, Subsampling (Sagawa et al, 2019(Sagawa et al, , 2020 from the machine learning literature, and a baseline RemoveToken that removes the treatment feature from input (see Supp C).…”
Section: Accuracy Of Feag Classifiersmentioning
confidence: 99%
“…For removing spurious correlations, a common principle underlying past work is to make a model's prediction invariant to the features that exhibit the correlation. This can be done by data augmentation (Kaushik et al, 2019), latent space removal (Ravfogel et al, 2020), subsampling (Sagawa et al, 2019(Sagawa et al, , 2020, or sample reweighing (Mahabadi et al, 2019;Orgad and Belinkov, 2022). In many cases, however, the correlated features may be important for the task and their complete removal can cause a degradation in task performance.…”
Section: Introductionmentioning
confidence: 99%