Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.262
|View full text |Cite
|
Sign up to set email alerts
|

How do Decisions Emerge across Layers in Neural Models? Interpretation with Differentiable Masking

Abstract: Attribution methods assess the contribution of inputs to the model prediction. One way to do so is erasure: a subset of inputs is considered irrelevant if it can be removed without affecting the prediction. Though conceptually simple, erasure's objective is intractable and approximate search remains expensive with modern deep NLP models. Erasure is also susceptible to the hindsight bias: the fact that an input can be dropped does not mean that the model 'knows' it can be dropped. The resulting pruning is over-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
33
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 38 publications
(39 citation statements)
references
References 35 publications
0
33
0
Order By: Relevance
“…Previous work on learning individual word masks only focuses on the first two properties De Cao et al, 2020). To satisfy the third property, We propose GMASK to implicitly detect word correlations and distribute the correlated words into a group (e.g.…”
Section: Explaining Models With Word Masksmentioning
confidence: 99%
“…Previous work on learning individual word masks only focuses on the first two properties De Cao et al, 2020). To satisfy the third property, We propose GMASK to implicitly detect word correlations and distribute the correlated words into a group (e.g.…”
Section: Explaining Models With Word Masksmentioning
confidence: 99%
“…where M l (P, H) and M Z l (P, H) ∈ R 3 are original output logits and output logits when applying the mask Z respectively. Compared to commonly used KL divergence (De Cao et al, 2020) or label equality (Feng et al, 2018), the euclidean distance between logits is a stricter constraint that narrows down the solution space and would lead to more faithful explanations 3 .…”
Section: Problem Formationmentioning
confidence: 99%
“…We select feature attribution baselines including co-attention itself, perturbation-based approaches LEAVEONEOUT (Li et al, 2016), LIME (Ribeiro et al, 2016), BACKSELECT (Carter et al, 2019), gradient-based approaches GRADIENT (Simonyan et al, 2014) and INTEGRATGRAD (Sundararajan et al, 2017) and a feature selection method DIFF-MASK (De Cao et al, 2020). The original DIFF-MASK is applied on text level, we derive an alignment variant for comparison in Appendix C.…”
Section: Baselinesmentioning
confidence: 99%
See 1 more Smart Citation
“…Another implementation for making word masks sparse is by adding L 0 regularization (Lei et al, 2016;Bastings et al, 2019;De Cao et al, 2020), while in the objective Equation 8, we regularize masks with a predefined prior distribution p 0 (R) as described in subsection 3.4.…”
Section: Connectionsmentioning
confidence: 99%