2019
DOI: 10.48550/arxiv.1902.05826
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Fairness of Risk Scores Beyond Classification: Bipartite Ranking and the xAUC Metric

Abstract: Where machine-learned predictive risk scores inform high-stakes decisions, such as bail and sentencing in criminal justice, fairness has been a serious concern. Recent work has characterized the disparate impact that such risk scores can have when used for a binary classification task. This may not account, however, for the more diverse downstream uses of risk scores and their non-binary nature. To better account for this, in this paper, we investigate the fairness of predictive risk scores from the point of v… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
7
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 24 publications
(26 reference statements)
1
7
0
Order By: Relevance
“…This pairwise ranking accuracy has a nice connection with the Mann-Whitney U-test [29], and aligns well with the equality gap metric [8] and the xAUC metric [24] for classification.…”
Section: Pairwise Ranking Accuracy As the Fairness Metricsupporting
confidence: 65%
See 2 more Smart Citations
“…This pairwise ranking accuracy has a nice connection with the Mann-Whitney U-test [29], and aligns well with the equality gap metric [8] and the xAUC metric [24] for classification.…”
Section: Pairwise Ranking Accuracy As the Fairness Metricsupporting
confidence: 65%
“…The composed system fairness metric is evaluated on the order (rank) of the product of the component scores. Evaluating the fairness metric on rank order aligns well with most real-world multi-model recommender systems, but can also be applied to classification [24,8,33].…”
Section: Introductionmentioning
confidence: 91%
See 1 more Smart Citation
“…There have also been two recent papers on pairwise fairness notions [4,19], both of which are focused on ranking with categorical groups and are methodologically different from us.…”
Section: Fair Rankingmentioning
confidence: 99%
“…Lastly, there are major experimental differences: they provide an in-depth case study of one real-world recommendation problem, whereas we provide a broad set of experiments on public and real-world data illustrating the effectiveness on both ranking and regression problems, for categorical or continuous protected attributes. [19] also provide pairwise fairness metrics based on AUC for bipartite ranking problems, but only consider categorical groups and propose a post-processing approach that fits a monotonic transform to an existing ranking model to optimize the proposed metrics. In contrast, we additionally handle regression problems and continuous protected attributes, and develop more flexible approaches that directly optimize for the desired pairwise fairness goals during training.…”
Section: Fair Rankingmentioning
confidence: 99%