2017
DOI: 10.1007/s10115-017-1116-3
|View full text |Cite
|
Sign up to set email alerts
|

Auditing black-box models for indirect influence

Abstract: Abstract-Data-trained predictive models see widespread use, but for the most part they are used as black boxes which output a prediction or score. It is therefore hard to acquire a deeper understanding of model behavior, and in particular how different features influence the model prediction. This is important when interpreting the behavior of complex models, or asserting that certain problematic attributes (like race or gender) are not unduly influencing decisions.In this paper, we present a technique for aud… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
36
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 137 publications
(46 citation statements)
references
References 30 publications
0
36
0
Order By: Relevance
“…Some methods transform the constrained optimization problem via the method of Lagrange multipliers [3,15,34,37,38,59,97,135,183,185] or add penalties to the objective [5,14,22,44,46,56,58,62,73,74,75,79,82,87,88,89,93,94,95,98,104,115,119,123,133,134,135,138,139,154,165,176,179,180,182], others use adversary techniques to maximize the system ability to predict the target while minimizing the ability to predict the sensitive attribute [189]. Post-Processing methods consist in transforming the model outputs in order to make them fair [2,7,…”
Section: Methods For Imposing Fairness In a Modelmentioning
confidence: 99%
See 2 more Smart Citations
“…Some methods transform the constrained optimization problem via the method of Lagrange multipliers [3,15,34,37,38,59,97,135,183,185] or add penalties to the objective [5,14,22,44,46,56,58,62,73,74,75,79,82,87,88,89,93,94,95,98,104,115,119,123,133,134,135,138,139,154,165,176,179,180,182], others use adversary techniques to maximize the system ability to predict the target while minimizing the ability to predict the sensitive attribute [189]. Post-Processing methods consist in transforming the model outputs in order to make them fair [2,7,…”
Section: Methods For Imposing Fairness In a Modelmentioning
confidence: 99%
“…Over the last few years, researchers have introduced a rich set of definitions formalizing different fairness desiderata that can be used for evaluating and designing ML systems [2,7,14,15,21,22,23,25,28,29,31,34,44,45,46,52,53,54,55,56,59,66,69,71,72,73,74,75,82,84,87,89,90,94,95,97,98,99,103,104,106,114,115,116,121,123,126,138,139,141,148,…”
Section: Causal Bayesian Network: An Essential Tool For Fairnessmentioning
confidence: 99%
See 1 more Smart Citation
“…Another related post hoc technique called black box auditing (Adler et al, 2016) can be used to decide the extent to which a specific feature contributes to the accuracy (percentage of correct predictions) of a trained model. To quantify the direct effect of a feature, we can replace the feature by random noise and see how much the model accuracy drops.…”
Section: Explaining What Data Were Fed Into the Modelmentioning
confidence: 99%
“…Such techniques use an interpretable model and apply it to the predictions returned by the black box (Hall, Phan, & Ambati, 2017). Currently many different explanation approaches exist, for example, providing logical statements (Lakkaraju, Bach, & Leskovec, 2016; Su, Wei, Varshney, & Malioutov, 2015; Wang et al, 2015; Wang & Rudin, 2014), local models (Rüping, 2005; Turner, 2016) or feature importance (Adler et al, 2018; Datta, Sen, & Zick, 2016; Goldstein, Kapelner, Bleich, & Pitkin, 2015; Puolamäki & Ukkonen, 2017).…”
Section: Theoretical Backgroundmentioning
confidence: 99%