2023
DOI: 10.48550/arxiv.2302.12578
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fairness in Language Models Beyond English: Gaps and Challenges

Abstract: With language models becoming increasingly ubiquitous, it has become essential to address their inequitable treatment of diverse demographic groups and factors. Most research on evaluating and mitigating fairness harms has been concentrated on English, while multilingual models and non-English languages have received comparatively little attention. This paper presents a survey of fairness in multilingual and non-English contexts, highlighting the shortcomings of current research and the difficulties faced by m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 81 publications
0
5
0
Order By: Relevance
“…As previously discussed, it is likely that documents written in English and from developed countries form the bulk of the training corpus -this may limit the nature of responses to specific queries and enhance existing biases. Therefore, there is an urgent need for culturally sensitive multi-lingual LLMs [40]. Moreover, in the current LLM landscape there is a lack of transparency around algorithm development and reporting related to decisions algorithms make during the review process.…”
Section: Ethical Considerationsmentioning
confidence: 99%
“…As previously discussed, it is likely that documents written in English and from developed countries form the bulk of the training corpus -this may limit the nature of responses to specific queries and enhance existing biases. Therefore, there is an urgent need for culturally sensitive multi-lingual LLMs [40]. Moreover, in the current LLM landscape there is a lack of transparency around algorithm development and reporting related to decisions algorithms make during the review process.…”
Section: Ethical Considerationsmentioning
confidence: 99%
“…While doing so, it is important to culturally contextualize NLP metrics and models. Instead of plainly translating English models into Bengali, Hindi, etc., we need to carefully consider the dimensions of fairness and types and sources of bias specific to that cultural context Ramesh et al, 2023). To address this gap, this paper proposes a methodology for developing culturally centered bias-evaluation datasets in NLP.…”
Section: Related Workmentioning
confidence: 99%
“…While doing so, it is important to culturally contextualize NLP metrics and models. Instead of plainly translating English models into Bengali, Hindi, etc., we need to carefully consider the dimensions of fairness and types and sources of bias specific to that cultural context (Malik et al, 2022;Ramesh et al, 2023). To address this gap, this paper proposes a methodology for developing culturally centered bias-evaluation datasets in NLP.…”
Section: Related Workmentioning
confidence: 99%