2022
DOI: 10.48550/arxiv.2205.09209
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

"I'm sorry to hear that": finding bias in language models with a holistic descriptor dataset

Abstract: As language models grow in popularity, their biases across all possible markers of demographic identity should be measured and addressed in order to avoid perpetuating existing societal harms. Many datasets for measuring bias currently exist, but they are restricted in their coverage of demographic axes, and are commonly used with preset bias tests that presuppose which types of biases the models exhibit. In this work, we present a new, more inclusive dataset, HOLISTICBIAS, which consists of nearly 600 descrip… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 33 publications
(49 reference statements)
0
4
0
Order By: Relevance
“…In open-ended language generation, prompts are often used to assess to what extent LMs yield undesirable output. Various benchmarks such as BOLD [17], HONEST [49], HolisticBias [61] and RealToxicityPrompts [24] exist for this purpose. Choenni et al [13] prompt language models to assess to what extent they have learnt stereotypes.…”
Section: Content Moderation In Language Modelsmentioning
confidence: 99%
“…In open-ended language generation, prompts are often used to assess to what extent LMs yield undesirable output. Various benchmarks such as BOLD [17], HONEST [49], HolisticBias [61] and RealToxicityPrompts [24] exist for this purpose. Choenni et al [13] prompt language models to assess to what extent they have learnt stereotypes.…”
Section: Content Moderation In Language Modelsmentioning
confidence: 99%
“…An additional study presents a large GBET dataset called HOLISTICBIAS for measuring bias. This dataset is assembled by using a set of demographic descriptor terms in a set of bias measurement templates and can be used to test bias in language models (Smith et al, 2022).…”
Section: Measuring Gender Biasmentioning
confidence: 99%
“…In addition to standard fairness evaluation datasets such as CrowS-Pairs (Nangia et al, 2020) 7 , and template-based fairness 7 Although many of the fairness metrics that are standard in NLP have flaws , we unfortunately have few alternatives. measurements such as the Word Embedding Association Test (WEAT) (Caliskan et al, 2017) and Sentence Encoder Association Test (SEAT), (May et al, 2019)), we also incorporate a new, larger bias measurement dataset, HolisticBias (HB), which was created with a combination of algorithmic and participatory processes to develop the most comprehensive descriptor term list available (Smith et al, 2022). We calculate the per-axis bias by measuring the fraction of pairs of descriptors in the HB dataset for which the distribution of pseudo-loglikelihoods (Nangia et al, 2020) in templated sentences significantly differs.…”
Section: Fairberta Is More Fairmentioning
confidence: 99%