Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society 2019
DOI: 10.1145/3306618.3314229
|View full text |Cite
|
Sign up to set email alerts
|

Faithful and Customizable Explanations of Black Box Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
201
0
4

Year Published

2019
2019
2023
2023

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 234 publications
(205 citation statements)
references
References 9 publications
0
201
0
4
Order By: Relevance
“…The explanation model is g : ℝ d ′ → ℝ, g ∈ G , where G is a class of potentially interpretable models, such as linear models, decision trees, or rule lists; given a model g ∈ G , it can be visualized as an explanation to the human expert (for details please refer to (Ribeiro, Singh, & Guestrin, )). Another example for a posthoc system is black box explanations through transparent approximations (BETA), a model‐agnostic framework for explaining the behavior of any black‐box classifier by simultaneously optimizing for fidelity to the original model and interpretability of the explanation introduced by Lakkaraju, Kamar, Caruana, and Leskovec ().…”
Section: General Approaches Of Explainable Ai Modelsmentioning
confidence: 99%
“…The explanation model is g : ℝ d ′ → ℝ, g ∈ G , where G is a class of potentially interpretable models, such as linear models, decision trees, or rule lists; given a model g ∈ G , it can be visualized as an explanation to the human expert (for details please refer to (Ribeiro, Singh, & Guestrin, )). Another example for a posthoc system is black box explanations through transparent approximations (BETA), a model‐agnostic framework for explaining the behavior of any black‐box classifier by simultaneously optimizing for fidelity to the original model and interpretability of the explanation introduced by Lakkaraju, Kamar, Caruana, and Leskovec ().…”
Section: General Approaches Of Explainable Ai Modelsmentioning
confidence: 99%
“…One-off explanations are still the most popular operationalisation of explainability algorithms [34], where the explainer outputs a one-size-fits-all explanation in an attempt to make the behaviour of a predictive system transparent. A slight improvement over this scenario is to enable the explainer to account for user preferences when generating the explanations [22,29], but this modality is not common either.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Akula et al [1] presented a dialogue-driven explainability system that uses contrastive explanations based on predictions derived from And-Or graphs and handcrafted ontology, however generalising this technique may be challenging as it requires hand-crafting separate ontology and And-Or graph for each application. Lakkaraju et al [22] introduced rule-based explanations that the user can personalise by choosing which features will appear in the explanation-an off-line personalisation. Google published their what-if tool 5 which provides the explainee with an interactive interface that allows generating contrastive explanations of selected data points by modifying their features, i.e., asking "What if?"…”
Section: Background and Related Workmentioning
confidence: 99%
“…Building on the growing fear of AI systems creating a "black box society" [19], work in the area of explainable AI has explored ways in which difficult to understand machine learning models may be represented and communicated more simply to users [20]. For example, model understanding through space explanations enables translating an arbitrary black-box representation of a machine learning system into decision sets which capture the behavior of the black box in specific circumstances [21]. In interactive machine learning, users can view and correct classifications made by a system [22].…”
Section: )mentioning
confidence: 99%