2021
DOI: 10.48550/arxiv.2106.08680
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Evaluating Gender Bias in Hindi-English Machine Translation

Abstract: With language models being deployed increasingly in the real world, it is essential to address the issue of the fairness of their outputs. The word embedding representations of these language models often implicitly draw unwanted associations that form a social bias within the model. The nature of gendered languages like Hindi, poses an additional problem to the quantification and mitigation of bias, owing to the change in the form of the words in the sentence, based on the gender of the subject. Additionally,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 2 publications
(2 reference statements)
0
1
0
Order By: Relevance
“…This is especially troubling for India, a pluralistic nation of 1.4 billion people, with fast-growing investments in NLP from the government 3 , and the private sector 4 . There is commendable recent work on NLP fairness in Indian languages like Hindi, Bengali, Telugu (Pujari et al, 2019;Malik et al, 2021;Gupta et al, 2021). But, for a nation with many religions, ethnicities, and cultures, recontextualizing NLP fairness needs to account for the various axes of social disparities in the Indian society, their proxies in language data, the disparate NLP capabilities in Indian languages, and the (lack of) resources for bias evaluation.…”
Section: Introductionmentioning
confidence: 99%
“…This is especially troubling for India, a pluralistic nation of 1.4 billion people, with fast-growing investments in NLP from the government 3 , and the private sector 4 . There is commendable recent work on NLP fairness in Indian languages like Hindi, Bengali, Telugu (Pujari et al, 2019;Malik et al, 2021;Gupta et al, 2021). But, for a nation with many religions, ethnicities, and cultures, recontextualizing NLP fairness needs to account for the various axes of social disparities in the Indian society, their proxies in language data, the disparate NLP capabilities in Indian languages, and the (lack of) resources for bias evaluation.…”
Section: Introductionmentioning
confidence: 99%