Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1523
|View full text |Cite
|
Sign up to set email alerts
|

Human-grounded Evaluations of Explanation Methods for Text Classification

Abstract: Due to the black-box nature of deep learning models, methods for explaining the models' results are crucial to gain trust from humans and support collaboration between AIs and humans. In this paper, we consider several model-agnostic and model-specific explanation methods for CNNs for text classification and conduct three human-grounded evaluations, focusing on different purposes of explanations:(1) revealing model behavior, (2) justifying model predictions, and (3) helping humans investigate uncertain predict… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
33
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 49 publications
(33 citation statements)
references
References 30 publications
0
33
0
Order By: Relevance
“…Another group of studies performs human evaluation of the outputs of explainability methods (Lertvittayakumjorn and Toni, 2019;Narayanan et al, 2018). Such studies exhibit low interannotator agreement and reflect mostly what appears to be reasonable and appealing to the annotators, not the actual properties of the method.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Another group of studies performs human evaluation of the outputs of explainability methods (Lertvittayakumjorn and Toni, 2019;Narayanan et al, 2018). Such studies exhibit low interannotator agreement and reflect mostly what appears to be reasonable and appealing to the annotators, not the actual properties of the method.…”
Section: Related Workmentioning
confidence: 99%
“…Existing studies for evaluating explainability heavily differ in their scope. Some concentrate on a single model architecture -BERT-LSTM (DeYoung et al, 2020), RNN (Arras et al, 2019), CNN (Lertvittayakumjorn and Toni, 2019), whereas a few consider more than one model (Guan et al, 2019;Poerner et al, 2018). Some studies concentrate on one particular dataset (Guan et al, 2019;Arras et al, 2019), while only a few generalize their findings over downstream tasks (DeYoung et al, 2020;Vashishth et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…Lipton (2018); Doshi-Velez and Kim (2017) and Rudin (2019) provide overviews on definitions and characterizations of interpretability. Lertvittayakumjorn and Toni (2019) classify three possible uses of text explanations: (i) revealing model behavior, (ii) justifying model predictions, and (iii) helping humans investigate uncertain predictions. Attempting to guarantee the faithfulness of a feature selection or explanation generation method is a more challenging question than finding explanations which humans find acceptable (Rudin, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…One of the advantages of GrASP lite is that it is an explainable model, making predictions based on rich and interpretable rules. These can be used to justify predictions, sometimes termed a local explanation (Lertvittayakumjorn and Toni, 2019) and also to understand the way the model works as a whole (termed global explanation), potentially enabling experts to build better classifiers.…”
Section: User Studymentioning
confidence: 99%