Proceedings of the First Workshop on Gender Bias in Natural Language Processing 2019
DOI: 10.18653/v1/w19-3823
|View full text |Cite
|
Sign up to set email alerts
|

Measuring Bias in Contextualized Word Representations

Abstract: Contextual word embeddings such as BERT have achieved state of the art performance in numerous NLP tasks. Since they are optimized to capture the statistical properties of training data, they tend to pick up on and amplify social stereotypes present in the data as well. In this study, we (1) propose a template-based method to quantify bias in BERT;(2) show that this method obtains more consistent results in capturing social biases than the traditional cosine based method; and (3) conduct a case study, evaluati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
275
1
1

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 280 publications
(321 citation statements)
references
References 25 publications
(38 reference statements)
4
275
1
1
Order By: Relevance
“…To the best of our knowledge, this is the first work to target affective dimensions in pre-trained contextualized word embeddings. Our findings are consistent with prior work suggesting that contextualized embeddings capture biases from training data (Zhao et al, 2019;Kurita et al, 2019) and that these models perform best when trained on in-domain data (Alsentzer et al, 2019).…”
Section: Related Worksupporting
confidence: 91%
“…To the best of our knowledge, this is the first work to target affective dimensions in pre-trained contextualized word embeddings. Our findings are consistent with prior work suggesting that contextualized embeddings capture biases from training data (Zhao et al, 2019;Kurita et al, 2019) and that these models perform best when trained on in-domain data (Alsentzer et al, 2019).…”
Section: Related Worksupporting
confidence: 91%
“…In this work, we examine the demographic parity, equality of opportunity for the positive class, and equality of opportunity for the negative class. 1 First, we demonstrate that there are significant differences in the log probability bias scores [36] of clinical text for different genders. These scores examine the probability of filling in the gender demographics given medical context.…”
Section: Introductionmentioning
confidence: 80%
“…debiasing) word embeddings. While individual works study how contextual word embeddings capture biases [5,36,62], to date the creation of debiasing methods has been limited to non-contextual word embeddings models (e.g. GLoVe [49], Word2Vec [45]).…”
Section: Fairness Of Word Embeddingsmentioning
confidence: 99%
See 2 more Smart Citations