Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-3002
|View full text |Cite
|
Sign up to set email alerts
|

Identifying and Reducing Gender Bias in Word-Level Language Models

Abstract: Many text corpora exhibit socially problematic biases, which can be propagated or amplified in the models trained on such data. For example, doctor cooccurs more frequently with male pronouns than female pronouns. In this study we (i) propose a metric to measure gender bias; (ii) measure bias in a text corpus and the text generated from a recurrent neural network language model trained on the text corpus; (iii) propose a regularization loss term for the language model that minimizes the projection of encoder-t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
112
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 143 publications
(113 citation statements)
references
References 16 publications
(17 reference statements)
1
112
0
Order By: Relevance
“…After training the baseline model, we implement our loss function and tune for the λ hyperparameter. We test the existing debiasing approaches, CDA and REG, as well but since Bordia and Bowman (2019) reported that results fluctuate substantially with different REG regularization coefficients, we perform hyperparameter tuning and report the best results in Table 2. Additionally, we implement a combination of our loss function and CDA and tune for λ.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…After training the baseline model, we implement our loss function and tune for the λ hyperparameter. We test the existing debiasing approaches, CDA and REG, as well but since Bordia and Bowman (2019) reported that results fluctuate substantially with different REG regularization coefficients, we perform hyperparameter tuning and report the best results in Table 2. Additionally, we implement a combination of our loss function and CDA and tune for λ.…”
Section: Methodsmentioning
confidence: 99%
“…We also implement the bias regularization method of Bordia and Bowman (2019) which debiases the word embedding during language model training by minimizing the projection of neutral words on the gender axis. We use hyperparameter tuning to find the best regularization coefficient and report results from the model trained with this coefficient.…”
Section: Existing Approachesmentioning
confidence: 99%
“…Zhao et al (2019) and Basta et al (2019) demonstrated gender bias in pretrained language modeling representations (ELMo), which translates into downstream tasks, but did not consider the language generated by the ELMo language model. Bordia and Bowman (2019), as well as Qian et al (2019) identified biases in a language modeling context and propose regularization strategies of generating certain words (e.g., "doctor") with differently gendered inputs.…”
Section: Background and Related Workmentioning
confidence: 99%
“…The study of biases in NLP systems is an active subfield. The majority of the work in the area is dedicated to pre-trained models, often via similaritybased analysis of the biases in input representations (Bolukbasi et al, 2016a;Garg et al, 2018;Chaloner and Maldonado, 2019;Bordia and Bowman, 2019;Tan and Celis, 2019;Zhao et al, , 2020, or an intermediate classification task (Recasens et al, 2013).…”
Section: Related Workmentioning
confidence: 99%