2022
DOI: 10.48550/arxiv.2202.01602
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Disagreement Problem in Explainable Machine Learning: A Practitioner's Perspective

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
56
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 39 publications
(57 citation statements)
references
References 0 publications
0
56
1
Order By: Relevance
“…However, the post hoc explainers we consider have also been shown to be inconsistent, unfaithful, and intractable [17,7,12,6,4]. Consequently, we believe that a potential source of negative societal impact in this work arises from practitioners overtrusting post hoc explainers [16,17]. While this is the case, our study demonstrates that the explainers backed with our proposed defense not only detect adversarial behavior but also faithfully identify the most important features in decisions.…”
Section: Discussionmentioning
confidence: 77%
See 1 more Smart Citation
“…However, the post hoc explainers we consider have also been shown to be inconsistent, unfaithful, and intractable [17,7,12,6,4]. Consequently, we believe that a potential source of negative societal impact in this work arises from practitioners overtrusting post hoc explainers [16,17]. While this is the case, our study demonstrates that the explainers backed with our proposed defense not only detect adversarial behavior but also faithfully identify the most important features in decisions.…”
Section: Discussionmentioning
confidence: 77%
“…We analyze and alleviate a single shortcoming of using post hoc explanations. However, the post hoc explainers we consider have also been shown to be inconsistent, unfaithful, and intractable [17,7,12,6,4]. Consequently, we believe that a potential source of negative societal impact in this work arises from practitioners overtrusting post hoc explainers [16,17].…”
Section: Discussionmentioning
confidence: 91%
“…Explanations should provide new insights: explanations should go beyond the "why, what, how" [20]. We must also be able to compare and contrast them, as explanations can disagree and directly contradict each other [21]. We extend these points and argue that developing metrics for explanations depends on both the audience as well as the capacity of the underlying explanation.…”
Section: Trusted But Not Trustworthy: a New Dark Patternmentioning
confidence: 82%
“…Lipton [65] examines the motivations underlying interest in interpretability, finding them to be diverse and occasionally discordant. Krishna et al [53] find that state-of-the-art explanation methods may disagree in terms of the explanations they output. Chandrasekaran et al [15] further conclude that existing explanations on VQA model do not actually make its responses and failures more predictable to a human.…”
Section: Limitations and Broader Impactmentioning
confidence: 99%