2022
DOI: 10.48550/arxiv.2201.11239
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Diagnosing AI Explanation Methods with Folk Concepts of Behavior

Abstract: When explaining AI behavior to humans, how is the communicated information being comprehended by the human explainee, and does it match what the explanation attempted to communicate? When can we say that an explanation is explaining something? We aim to provide an answer by leveraging theory of mind literature about the folk concepts that humans use to understand behavior.We establish a framework of social attribution by the human explainee, which describes the function of explanations: the concrete informatio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 81 publications
(111 reference statements)
0
1
0
Order By: Relevance
“…ϕ f should be continuous. This robustness can be seen as a measure of coherence [18]. his type of robustness appears to be difficult to achieve, however, because a sequence of empirical counterexamples have been found in which several ethods like LIME, SHAP and Integrated Gradients are not robust [14,37,10,3].…”
Section: Attribution Methods and Counterfactual Explanationsmentioning
confidence: 99%
“…ϕ f should be continuous. This robustness can be seen as a measure of coherence [18]. his type of robustness appears to be difficult to achieve, however, because a sequence of empirical counterexamples have been found in which several ethods like LIME, SHAP and Integrated Gradients are not robust [14,37,10,3].…”
Section: Attribution Methods and Counterfactual Explanationsmentioning
confidence: 99%