2019
DOI: 10.1007/978-3-030-29908-8_4
|View full text |Cite
|
Sign up to set email alerts
|

Explaining Deep Learning Models with Constrained Adversarial Examples

Abstract: Machine learning algorithms generally suffer from a problem of explainability. Given a classification result from a model, it is typically hard to determine what caused the decision to be made, and to give an informative explanation. We explore a new method of generating counterfactual explanations, which instead of explaining why a particular classification was made explain how a different outcome can be achieved. This gives the recipients of the explanation a better way to understand the outcome, and provide… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 29 publications
(33 citation statements)
references
References 8 publications
(8 reference statements)
0
31
0
Order By: Relevance
“…The researchers suggest the use of the Manhattan distance weighted by the inverse median absolute deviation to calculate the proximity of a counterfactual to the input data example. Another case of counterfactual explanation generation regarded as an optimization problem is the ''Constrained Adversarial Examples'' framework [148]. Adversarial examples that could serve as the basis for the counterfactual explanation of the output of deep learning models are searched for with the aim of minimizing the loss with respect to the attributes (features) between the original and counterfactual data examples.…”
Section: ) Explainability Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…The researchers suggest the use of the Manhattan distance weighted by the inverse median absolute deviation to calculate the proximity of a counterfactual to the input data example. Another case of counterfactual explanation generation regarded as an optimization problem is the ''Constrained Adversarial Examples'' framework [148]. Adversarial examples that could serve as the basis for the counterfactual explanation of the output of deep learning models are searched for with the aim of minimizing the loss with respect to the attributes (features) between the original and counterfactual data examples.…”
Section: ) Explainability Methodsmentioning
confidence: 99%
“…Indeed, contfactuals are particularly suitable for informing the end-user why a given data example is assigned a particular class label. Thus, the outlined classification-oriented frameworks are evaluated on classifiers based on logistic regression [55], [136], [153], [158], decision trees [46], [80], [122], [140], [150], [155], [159], gradient boosted decision trees [147], support vector machines [131], [138], [146], random forests [81], [86], [142]- [144], neural networks [6], [48], [49], [91], [129], [130], [133], [135], [139], [141], [145], [148], [151], or combinations of these [100], [105], [134], [152], [154], [160]. In three studies [67], [128], [137], the classifiers used in the experiments are not specified.…”
Section: ) Ai Problemmentioning
confidence: 99%
See 1 more Smart Citation
“…The existing CE methods can be categorized into gradient-based [21,33], autoencoder [5,19], SAT [16], or mixedinteger linear optimization (MILO) [4,15,29,32]. Since our cost function is non-differentiable due to the discrete nature of a permutation σ over features, we focus on MILO-based methods, which can directly handle such functions.…”
Section: Related Workmentioning
confidence: 99%
“…Constraint 13, if H is a LM, Constraint (14)(15)(16)(17), if H is a TE, Constraint (18)(19)(20)(21), if H is a MLP,…”
Section: Overall Formulationmentioning
confidence: 99%