2022
DOI: 10.48550/arxiv.2201.08542
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Can Model Compression Improve NLP Fairness

Abstract: Model compression techniques are receiving increasing attention; however, the effect of compression on model fairness is still under explored. This is the first paper to examine the effect of distillation and pruning on the toxicity and bias of generative language models. We test Knowledge Distillation and Pruning methods on the GPT2 model and found a consistent pattern of toxicity and bias reduction after model distillation; this result can be potentially interpreted by existing line of research which describ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
3
1

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 14 publications
0
3
1
Order By: Relevance
“…Specifically regarding bias in the distillation or model compression setting, Xu and Hu (2022) report reduction in bias in contrast to our findings, although in a generation application. However, Gupta et al (2022) makes clear that biases from the training data can also be preserved or exacerbated in a similar distillation setting.…”
Section: Related Workcontrasting
confidence: 99%
“…Specifically regarding bias in the distillation or model compression setting, Xu and Hu (2022) report reduction in bias in contrast to our findings, although in a generation application. However, Gupta et al (2022) makes clear that biases from the training data can also be preserved or exacerbated in a similar distillation setting.…”
Section: Related Workcontrasting
confidence: 99%
“…Additionally, [8] presented a regularisation procedure that aims at debiasing a language model by minimising the projection of encoder-trained embeddings onto a subspace that encodes gender. Similarly, [59] used model compression techniques, a type of regularisation techniques, to reduce toxicity and bias originally present in generative language models. The system proposed by [32] mitigates bias by employing counterfactual data augmentation, proving that modifying the training data works better than changing the actual geometry of the embeddings.…”
Section: Related Workmentioning
confidence: 99%
“…Xu and Hu 37 showed that there is reduction of toxicity and bias in compressed GPT2 model compressed via Knowledge Distillation. In computer vision models, Quantization and Pruning methods have compromised fairness because performance of samples with under-represented features is sacrificed after compression ( 38 , 39 ).…”
Section: Bias In Compressed Modelsmentioning
confidence: 99%