2018
DOI: 10.1007/978-3-319-91473-2_9
|View full text |Cite
|
Sign up to set email alerts
|

Comparison-Based Inverse Classification for Interpretability in Machine Learning

Abstract: In the context of post-hoc interpretability, this paper addresses the task of explaining the prediction of a classifier, considering the case where no information is available, neither on the classifier itself, nor on the processed data (neither the training nor the test data). It proposes an instance-based approach whose principle consists in determining the minimal changes needed to alter a prediction: given a data point whose classification must be explained, the proposed method consists in identifying a cl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
68
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
5
1
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 102 publications
(78 citation statements)
references
References 20 publications
(10 reference statements)
2
68
0
Order By: Relevance
“…The latter usually exploit the internal structure of the underlying ML model, such as the trained weights of a neural network, while the former are based on general principles which work for arbitrary ML models -often by only assuming access to the prediction function of an already fitted model. Several model-agnostic counterfactual methods have been proposed [8,11,16,18,25,29,37]. Apart from Grath et al [11], these approaches are limited to classification.…”
Section: Related Workmentioning
confidence: 99%
“…The latter usually exploit the internal structure of the underlying ML model, such as the trained weights of a neural network, while the former are based on general principles which work for arbitrary ML models -often by only assuming access to the prediction function of an already fitted model. Several model-agnostic counterfactual methods have been proposed [8,11,16,18,25,29,37]. Apart from Grath et al [11], these approaches are limited to classification.…”
Section: Related Workmentioning
confidence: 99%
“…Subject to our experimental conditions, approximately 10 min were required for path planning per instance. A more practical path planning can be expected by combining our framework with techniques for finding intervention points, such as counterfactual explanations [41][42][43][44][45] .…”
Section: Discussionmentioning
confidence: 99%
“…Therefore, some authors find counterfactuals by applying ''smart'' iterative procedures in the input space [94] or search the counterfactual in a (growing) neighborhood of the starting point x; the neighborhood is enlarged until a counterfactual that allows to obtain the desired output is obtained [95]. Of note, some authors provide classifier interpretations that, though opposite to counterfactual explanations, allow anyway to interpret the classifier and provide counterfactuals.…”
Section: ) Explaining Prediction and Providing Suggestions With Counmentioning
confidence: 99%
“…In our work, after predicting the fitness score for a novel measurement x, the DT computes counterfactuals in a way similar to that used in [95].…”
Section: B Treating Novel Measurements To Provide Suggestions Based mentioning
confidence: 99%