2019
DOI: 10.48550/arxiv.1901.04592
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Interpretable machine learning: definitions, methods, and applications

W. James Murdoch,
Chandan Singh,
Karl Kumbier
et al.

Abstract: Machine-learning models have demonstrated great success in learning complex patterns that enable them to make predictions about unobserved data. In addition to using models for prediction, the ability to interpret what a model has learned is receiving an increasing amount of attention. However, this increased focus has led to considerable confusion about the notion of interpretability. In particular, it is unclear how the wide array of proposed interpretation methods are related, and what common concepts can b… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
121
0
5

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 90 publications
(126 citation statements)
references
References 66 publications
0
121
0
5
Order By: Relevance
“…Explainability and interpretability are topics of growing interest in the machine learning community [Ribeiro et al, 2016, Lundberg and Lee, 2017, Adadi and Berrada, 2018, Rudin, 2019, Murdoch et al, 2019, Molnar, 2020. While there has been some focus on what Dasgupta et al [2020] calls postmodeling explainability, or the ability to explain the output of a black-box model [Ribeiro et al, 2016, Lundberg and Lee, 2017, Kauffmann et al, 2019, the practice has also been criticized in contrast with pre-modelling explainability, or the use of interpretable models to begin with [Rudin, 2019].…”
Section: Preliminaries and Problem Definitionmentioning
confidence: 99%
“…Explainability and interpretability are topics of growing interest in the machine learning community [Ribeiro et al, 2016, Lundberg and Lee, 2017, Adadi and Berrada, 2018, Rudin, 2019, Murdoch et al, 2019, Molnar, 2020. While there has been some focus on what Dasgupta et al [2020] calls postmodeling explainability, or the ability to explain the output of a black-box model [Ribeiro et al, 2016, Lundberg and Lee, 2017, Kauffmann et al, 2019, the practice has also been criticized in contrast with pre-modelling explainability, or the use of interpretable models to begin with [Rudin, 2019].…”
Section: Preliminaries and Problem Definitionmentioning
confidence: 99%
“…Deep learning systems have been adopted in many areas from medicine to autonomous driving (Ahmad et al, 2018;Claybrook and Kildare, 2018) and as these algorithms are incorporated, the need for explainable and transparent models becomes more urgent. One approach researchers use to overcome the inherent ambiguity of these black-box methods is to develop additional models to learn and explain the decisions of existing models, analyze when these models fail, and introduce a human-in-the-loop component to improve performance (Ribeiro et al, 2016;Murdoch et al, 2019;Poursabzi-Sangdeh et al, 2018). Another way to tackle this challenge is to design models with a goal of interpretability in place when development starts (Ridgeway et al, 1998;Rudin, 2018;Gilpin et al, 2018;Lahav et al, 2018;Hooker et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…While post-modeling explainability focuses on giving reasoning behind decisions made by black box models, pre-modeling explainability deals with ML systems that are inherently understandable or perceivable by humans. One of the canonical approaches to pre-modelling explainability builds on decision trees [35,37]. In fact, a significant amount of work on explainable clustering is based on unsupervised decision trees [3,17,20,21,29,36].…”
Section: Introductionmentioning
confidence: 99%