2019
DOI: 10.48550/arxiv.1904.00045
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Interpreting Black Box Models via Hypothesis Testing

Collin Burns,
Jesse Thomason,
Wesley Tansey

Abstract: While many methods for interpreting machine learning models have been proposed, they are often ad hoc, difficult to interpret, and come with limited guarantees. This is especially problematic in science and medicine, where model interpretations may be reported as discoveries or guide patient treatments. As a step toward more principled and reliable interpretations, in this paper we reframe black box model interpretability as a multiple hypothesis testing problem. The task is to discover "important" features by… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 16 publications
(23 reference statements)
0
7
0
Order By: Relevance
“…4) Randomization and Feature Testing: The Interpretability Randomization Test (IRT) and the One-Shot Feature Test (OSFT) introduced by Burns et al [55] focuses on discovering important features by replacing the features with uninformative counterfactuals. Modeling the feature replacement with a hypothesis testing framework, the authors illustrate an interesting way to examine contextual importance.…”
Section: A Perturbation-basedmentioning
confidence: 99%
“…4) Randomization and Feature Testing: The Interpretability Randomization Test (IRT) and the One-Shot Feature Test (OSFT) introduced by Burns et al [55] focuses on discovering important features by replacing the features with uninformative counterfactuals. Modeling the feature replacement with a hypothesis testing framework, the authors illustrate an interesting way to examine contextual importance.…”
Section: A Perturbation-basedmentioning
confidence: 99%
“…In contrast to LIME which only locally explains, the SHAP method can be used for global explanation. There are also alternative methods which include influence functions [32], a multiple hypothesis testing framework [33], and many other methods.…”
Section: Related Workmentioning
confidence: 99%
“…Another very popular method is the SHAP [49] which takes a game-theoretic approach for optimizing a regression loss function based on Shapley values [33]. Alternative methods are influence functions [31], a multiple hypothesis testing framework [9], and many other methods.…”
Section: Related Workmentioning
confidence: 99%