2022
DOI: 10.48550/arxiv.2206.08454
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Quantifying Feature Contributions to Overall Disparity Using Information Theory

Abstract: When a machine-learning algorithm makes biased decisions, it can be helpful to understand the "sources" of disparity to explain why the bias exists. Towards this, we examine the problem of quantifying the contribution of each individual feature to the observed disparity. If we have access to the decision-making model, one potential approach (inspired from intervention-based approaches in explainability literature) is to vary each individual feature (while keeping the others fixed), and use the resulting change… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 34 publications
0
5
0
Order By: Relevance
“…However, this calculation will not be accurate and outcomes can significantly differ from true model outcomes i.e. when feature đť‘Ž is not used to train the model [13,19,24,33]. In fact, (interventional) SHAP simulates the removal of features by marginalising over their marginal distributions and not by re-training a new model without such features [36].…”
Section: Refresh: Theory and Methodsmentioning
confidence: 99%
“…However, this calculation will not be accurate and outcomes can significantly differ from true model outcomes i.e. when feature đť‘Ž is not used to train the model [13,19,24,33]. In fact, (interventional) SHAP simulates the removal of features by marginalising over their marginal distributions and not by re-training a new model without such features [36].…”
Section: Refresh: Theory and Methodsmentioning
confidence: 99%
“…In Section 3, we first introduce the problem setup for quantifying non-exempt disparity as discussed in [8,29], and present several canonical examples and candidate measures, understanding their pros and cons, until we arrive at the proposed measure in [8] that satisfies the desirable properties. In Section 4, we review how PID can help in assessing the contributions of either features or data points with applications in feature selection (as discussed in [31]). Related works include [35][36][37][38].…”
Section: Scenario 3: Formalizing Tradeoffs In Distributed Environmentsmentioning
confidence: 99%
“…Towards answering this question, ref. [31] proposes two measures for quantifying the contribution of each feature to the overall disparity. The first measure, which is referred to as interventional contribution, is defined as follows:…”
Section: Information-theoretic Measuresmentioning
confidence: 99%
See 2 more Smart Citations