2018
DOI: 10.1073/pnas.1720347115
|View full text |Cite
|
Sign up to set email alerts
|

Word embeddings quantify 100 years of gender and ethnic stereotypes

Abstract: Word embeddings are a powerful machine-learning framework that represents each English word by a vector. The geometric relationship between these vectors captures meaningful semantic relationships between the corresponding words. In this paper, we develop a framework to demonstrate how the temporal dynamics of the embedding helps to quantify changes in stereotypes and attitudes toward women and ethnic minorities in the 20th and 21st centuries in the United States. We integrate word embeddings trained on 100 y … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

21
861
1
6

Year Published

2018
2018
2023
2023

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 752 publications
(896 citation statements)
references
References 38 publications
21
861
1
6
Order By: Relevance
“…We also explore the stability of embeddings, but focus on a broader range of factors, and consider the effect of stability on downstream tasks. In contrast, Antoniak and Mimno focus on using word embeddings to analyze language (e.g., Garg et al, 2018), rather than to perform tasks.…”
Section: Related Workmentioning
confidence: 99%
“…We also explore the stability of embeddings, but focus on a broader range of factors, and consider the effect of stability on downstream tasks. In contrast, Antoniak and Mimno focus on using word embeddings to analyze language (e.g., Garg et al, 2018), rather than to perform tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Natural language processing (NLP) algorithms have been reported to incorporate inherent bias when trained on human language . NLP techniques such as word embedding are now used to objectively evaluate gender and ethnic stereotypes in text data . In recent years, there have been unfortunate examples of nonmedical NLP and machine learning algorithms that have produced biased recommendations .…”
Section: Introductionmentioning
confidence: 99%
“…1,2 NLP techniques such as word embedding are now used to objectively evaluate gender and ethnic stereotypes in text data. 3 In recent years, there have been unfortunate examples of nonmedical NLP and machine learning algorithms that have produced biased recommendations. 4 These setbacks risk jeopardizing physician trust in machine learning-based clinical decision support tools.…”
Section: Introductionmentioning
confidence: 99%
“…For example, studies have shown that-based on vectors learned from text corpora-computer programmer is more to male than female, and further found that this type of analogy reflects societal bias, which can be quantified through word vectors (Garg et al, 2018).…”
Section: Lsmentioning
confidence: 99%