Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.209
|View full text |Cite
|
Sign up to set email alerts
|

Type B Reflexivization as an Unambiguous Testbed for Multilingual Multi-Task Gender Bias

Abstract: The one-sided focus on English in previous studies of gender bias in NLP misses out on opportunities in other languages: English challenge datasets such as GAP and Wino-Gender highlight model preferences that are "hallucinatory", e.g., disambiguating genderambiguous occurrences of 'doctor' as male doctors. We show that for languages with type B reflexivization, e.g., Swedish and Russian, we can construct multi-task challenge datasets for detecting gender bias that lead to unambiguously wrong model predictions:… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 15 publications
(14 citation statements)
references
References 22 publications
(22 reference statements)
0
12
0
Order By: Relevance
“…Several works created synthetic datasets to evaluate gender bias (Kiritchenko and Mohammad, 2018;González et al, 2020;Renduchintala and Williams, 2021), e.g., in the context of coreference (Rudinger et al, 2017;Zhao et al, 2018) and machine translation (Stanovsky et al, 2019;Prates et al, 2019;Kocmi et al, 2020), and some works used synthetic datasets to debias models (Saunders et al, 2020;Zhao et al, 2018). Webster et al (2018) and Gonen and Webster (2020), collected natural medium-scale (4.4K sentences) datasets from Wikipedia and reddit, re-spectively, and use them to evaluate gender bias in models of coreference resolution and machine translation.…”
Section: Related Workmentioning
confidence: 99%
“…Several works created synthetic datasets to evaluate gender bias (Kiritchenko and Mohammad, 2018;González et al, 2020;Renduchintala and Williams, 2021), e.g., in the context of coreference (Rudinger et al, 2017;Zhao et al, 2018) and machine translation (Stanovsky et al, 2019;Prates et al, 2019;Kocmi et al, 2020), and some works used synthetic datasets to debias models (Saunders et al, 2020;Zhao et al, 2018). Webster et al (2018) and Gonen and Webster (2020), collected natural medium-scale (4.4K sentences) datasets from Wikipedia and reddit, re-spectively, and use them to evaluate gender bias in models of coreference resolution and machine translation.…”
Section: Related Workmentioning
confidence: 99%
“…Danish has gendered possessive pronouns, but nongendered reflexive pronouns. This has made it useful as an unambiguous testbed for gender bias in natural language inference models, machine translation models, and language models (González et al, 2020). But automatic coreference resolution for Danish has received no attention, and there was no established evaluation set for this task.…”
Section: Related Workmentioning
confidence: 99%
“…With the rise in the economy comes more slots in schools, but still, the money is needed, and with the economy rising, some families can't afford to pay the fees for all their children. Reports show that even though most families can send their children to elementary, primary and junior high school, they cannot send them to tertiary schools (González, et al, 2020). Another factor affecting gender inequality in education is the number of children in the household.…”
Section: Rural Areas and Gender Bias In Educationmentioning
confidence: 99%