“…Text data reflects the social and cultural biases in the world, and NLP models and applications trained on such data have been shown to reproduce and amplify those biases. Discrimination has been identified across diverse sensitive attributes including gender, disability, race, and religion (Caliskan et al, 2017;May et al, 2019;Garimella et al, 2019;Nangia et al, 2020;Li et al, 2020). While early work focused on debiasing typically binarized protected attributes in isolation (e.g., age, gender, or race; Caliskan et al (2017)), more recent work has adopted a more realistic scenario with multiple sensitive attributes (Li et al, 2018) or attributes covering several classes (Manzini et al, 2019).…”