Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence 2019
DOI: 10.24963/ijcai.2019/814
|View full text |Cite
|
Sign up to set email alerts
|

mdfa: Multi-Differential Fairness Auditor for Black Box Classifiers

Abstract: Machine learning algorithms are increasingly involved in sensitive decision-making process with adversarial implications on individuals. This paper presents mdfa, an approach that identifies the characteristics of the victims of a classifier's discrimination. We measure discrimination as a violation of multi-differential fairness. Multi-differential fairness is a guarantee that a black box classifier's outcomes do not leak information on the sensitive attributes of a small group of individuals. We reduce the p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
10
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(10 citation statements)
references
References 11 publications
0
10
0
Order By: Relevance
“…The work allows for black-box type algorithms to be audited for bias mitigating injustice, but also understanding the treatment. It further reinforces that injustices can be identified by performing a comparison between the treatment received by one group versus another, through the MDFA, which aims to ensure that minority groups receive proper treatment regardless of their sensitive attributes [24].…”
Section: Mitigation Techniques and Modelsmentioning
confidence: 85%
See 4 more Smart Citations
“…The work allows for black-box type algorithms to be audited for bias mitigating injustice, but also understanding the treatment. It further reinforces that injustices can be identified by performing a comparison between the treatment received by one group versus another, through the MDFA, which aims to ensure that minority groups receive proper treatment regardless of their sensitive attributes [24].…”
Section: Mitigation Techniques and Modelsmentioning
confidence: 85%
“…Except for [24] and [8], the model proposals were primarily white-box classification. The former proposes a model for bias elimination using Multi-Differential Fairness by integrating in-processing and post-processing, whereas the latter proposes that the focus of algorithm transparency should be on the output rather than the whole decision-making process of the algorithm.…”
Section: Discussionmentioning
confidence: 99%
See 3 more Smart Citations