The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2022
DOI: 10.48550/arxiv.2203.11933
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Prompt Array Keeps the Bias Away: Debiasing Vision-Language Models with Adversarial Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 0 publications
0
5
0
Order By: Relevance
“…Wang et al (2021) propose to remove dimensions in the CLIP embedding that are highly correlated with gender attributes. Berg et al (2022) debias the CLIP models with prompt learning via an adversarial approach. Recently, Zhang & Ré (2022) address the group robustness of vision-language models with contrastive learning.…”
Section: Biases In Vision Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…Wang et al (2021) propose to remove dimensions in the CLIP embedding that are highly correlated with gender attributes. Berg et al (2022) debias the CLIP models with prompt learning via an adversarial approach. Recently, Zhang & Ré (2022) address the group robustness of vision-language models with contrastive learning.…”
Section: Biases In Vision Modelsmentioning
confidence: 99%
“…Prompts for text-image retrieval on FairFace Dataset. We adopt the 10 training concepts from (Berg et al, 2022) to construct the prompts for FairFace. These concepts are irrelevant to gender, race, or age, which makes them suitable for evaluating the model biases.…”
Section: B2 Lemma 42mentioning
confidence: 99%
See 1 more Smart Citation
“…Several methodologies to measure and mitigate bias cannot be applied in our setting given the lack of public access to GPT-3's model architecture or training dataset, and the enormous resources needed to retrain the model from scratch. In particular, this includes training data augmentation (Sen et al, 2021), adjusting model behaviour via adversarial learning (Zhang et al, 2018;Berg et al, 2022), and amending model embeddings (Dev and Phillips, 2019).…”
Section: Measuring Biasmentioning
confidence: 99%
“…Several methodologies to measure and mitigate bias cannot be applied in our setting given the lack of public access to GPT-3's model architecture or training dataset, and the enormous resources needed to retrain the model from scratch. In particular, this includes training data augmentation (Sen et al, 2021), adjusting model behaviour via adversarial learning Berg et al, 2022), and amending model embeddings . Our analysis instead focuses on the text-level bias of model-generated outputs which we measure via a composite score based on the prevalence of certain gender-laden terms, and debiasing methods which require no access to the model architecture, nor original training data.…”
Section: Measuring Biasmentioning
confidence: 99%