Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency 2021
DOI: 10.1145/3442188.3445883
|View full text |Cite
|
Sign up to set email alerts
|

Removing Spurious Features can Hurt Accuracy and Affect Groups Disproportionately

Abstract: The presence of spurious features interferes with the goal of obtaining robust models that perform well across many groups within the population. A natural remedy is to remove spurious features from the model. However, in this work we show that removal of spurious features can decrease accuracy due to the inductive biases of overparameterized models. We completely characterize how the removal of spurious features affects accuracy across different groups (more generally, test distributions) in noiseless overpar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
24
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 34 publications
(25 citation statements)
references
References 31 publications
1
24
0
Order By: Relevance
“…In addition, while a variety of recent papers have proposed methods for removing spurious cues from training data or ''de-biasing'' models, recent work has shown that this can be damaging for model accuracy. 88 In contrast to a focus on statistical properties of datasets as a site for addressing and mitigating harms, Denton et al 89 propose a research agenda in the ''data genealogy'' paradigm that promotes critical assessment of the design choices with respect to the data sources, theoretical motivations, and methods used for constructing datasets. Prospective accounting for dataset contents using some of the methods discussed at the end of the previous section can offset the potential of post-hoc documentation debt that can be incurred otherwise.…”
Section: Introspectionmentioning
confidence: 99%
“…In addition, while a variety of recent papers have proposed methods for removing spurious cues from training data or ''de-biasing'' models, recent work has shown that this can be damaging for model accuracy. 88 In contrast to a focus on statistical properties of datasets as a site for addressing and mitigating harms, Denton et al 89 propose a research agenda in the ''data genealogy'' paradigm that promotes critical assessment of the design choices with respect to the data sources, theoretical motivations, and methods used for constructing datasets. Prospective accounting for dataset contents using some of the methods discussed at the end of the previous section can offset the potential of post-hoc documentation debt that can be incurred otherwise.…”
Section: Introspectionmentioning
confidence: 99%
“…Recently, Vision Transformers (ViTs, Dosovitskiy et al (2021)) sparked great interest in the literature, as a radically new model architecture offering significant accuracy improvements and with hope of new robustness benefits. Over the past decade, there has been extensive work on understanding the robustness of convolution-based neural architectures, as the dominant design for visual tasks; researchers have explored adversarial robustness (Szegedy et al, 2013), domain generalization (Xiao et al, 2021;Khani & Liang, 2021), feature biases (Brendel & Bethge, 2019;Geirhos et al, 2018;Hermann et al, 2020). As a result, with the new promise of vision transformers, it is critical to understand their properties and in particular their robustness.…”
Section: Introductionmentioning
confidence: 99%
“…Fairness-oriented [133,80,59,79] Robustness-oriented [144,72,136] Equal Mergers [110,112,132,122] Table 1: Taxonomy of data collection and quality techniques for deep learning.…”
Section: Data Discoverymentioning
confidence: 99%
“…After training, the classifier's output becomes independent of the sensitive group. Fair training without spurious features [72] addresses the problem of preventing feature removal from being discriminating. A self-training technique is proposed to mitigate accuracy degradation and biased effects (Figure 32).…”
Section: Robustness-oriented Approachesmentioning
confidence: 99%