Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.149
|View full text |Cite
|
Sign up to set email alerts
|

A Survey of Race, Racism, and Anti-Racism in NLP

Abstract: Despite inextricable ties between race and language, little work has considered race in NLP research and development. In this work, we survey 79 papers from the ACL anthology that mention race. These papers reveal various types of race-related bias in all stages of NLP model development, highlighting the need for proactive consideration of how NLP systems can uphold racial hierarchies. However, persistent gaps in research on race and NLP remain: race has been siloed as a niche topic and remains ignored in many… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
39
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
5

Relationship

1
9

Authors

Journals

citations
Cited by 41 publications
(40 citation statements)
references
References 108 publications
1
39
0
Order By: Relevance
“…We make use of pretrained language models to both generate and retrieve text in this work. Representations from pretrained language models are known to cause ethical concerns, such as perpetuating racial or gender bias (Field et al, 2021;Gala et al, 2020). We advise using caution and adopting a post-processing strategy to filter potentially offensive text produced by pretrained language models before releasing text content to users.…”
Section: Ethical Considerationsmentioning
confidence: 99%
“…We make use of pretrained language models to both generate and retrieve text in this work. Representations from pretrained language models are known to cause ethical concerns, such as perpetuating racial or gender bias (Field et al, 2021;Gala et al, 2020). We advise using caution and adopting a post-processing strategy to filter potentially offensive text produced by pretrained language models before releasing text content to users.…”
Section: Ethical Considerationsmentioning
confidence: 99%
“…Unsurprisingly, gendered and racial disparities have been documented in a number of language technologies [37,79,141], and processes of creating resources and technologies may further entrench such disparities [25,38,144]. For more detail see [53].…”
Section: Social Context: Social Variation and Language Discriminationmentioning
confidence: 99%
“…The risks of models replicating or worsening harmful biases may grow as we train on ever larger data samples (Bender et al, 2021). Training models on data with representational issues can lead them to treat particular demographic groups unfairly and/or poorly (Barocas et al, 2017;Mehrabi et al, 2021), a problem that is particularly egregious for historically marginalized groups, including people of color (Field et al, 2021), and women (Hendricks et al, 2018). For example, models learned the stereotype that "women like shopping" when they were trained on data where most or all of shoppers are women, and they learned * Equal contribution.…”
Section: Introductionmentioning
confidence: 99%