Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP 2019
DOI: 10.18653/v1/w19-4813
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating Recurrent Neural Network Explanations

Abstract: Recently, several methods have been proposed to explain the predictions of recurrent neural networks (RNNs), in particular of LSTMs. The goal of these methods is to understand the network's decisions by assigning to each input variable, e.g., a word, a relevance indicating to which extent it contributed to a particular prediction. In previous works, some of these methods were not yet compared to one another, or were evaluated only qualitatively. We close this gap by systematically and quantitatively comparing … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
49
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2
2

Relationship

2
5

Authors

Journals

citations
Cited by 57 publications
(50 citation statements)
references
References 43 publications
1
49
0
Order By: Relevance
“…A recent study [8] proposes to objectively evaluate explanation for sequential data using ground truth information in a toy task. The idea of this evaluation metric is to add or subtract two numbers within an input sequence and measure the correlation between the relevances assigned to the elements of the sequence and the two input numbers.…”
Section: Evaluating Quality Of Explanationsmentioning
confidence: 99%
“…A recent study [8] proposes to objectively evaluate explanation for sequential data using ground truth information in a toy task. The idea of this evaluation metric is to add or subtract two numbers within an input sequence and measure the correlation between the relevances assigned to the elements of the sequence and the two input numbers.…”
Section: Evaluating Quality Of Explanationsmentioning
confidence: 99%
“…In this section we discuss various input saliency methods for NLP as alternatives to attention: gradient-based ( §3.1), propagation-based ( §3.2), and occlusion-based methods ( §3.3), following Arras et al (2019). We do not endorse any specific method 1 , but rather try to give an overview of methods and how they differ.…”
Section: Saliency Methodsmentioning
confidence: 99%
“…Relevance is redistributed until we arrive at the input layers. While LRP requires implementing a custom backward pass, it does allow precise control to preserve relevance, and it has been shown to work better than using gradient-based methods on text classification (Arras et al, 2019).…”
Section: Propagation-based Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…For example, Zhang et al (2018) propose the pointing game task, in which the highest-relevance pixel for an image classifier input must belong to the object described by the target output class. Within this framework, ), Poerner et al (2018, Arras et al (2019), and Yang and Kim (2019) construct datasets in which input features exhibit experimentally controlled notions of importance, yielding "ground truth" attributions against which heatmaps can be evaluated.…”
Section: Related Workmentioning
confidence: 99%