Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.347
|View full text |Cite
|
Sign up to set email alerts
|

Learning Variational Word Masks to Improve the Interpretability of Neural Text Classifiers

Abstract: To build an interpretable neural text classifier, most of the prior work has focused on designing inherently interpretable models or finding faithful explanations. A new line of work on improving model interpretability has just started, and many existing methods require either prior information or human annotations as additional inputs in training. To address this limitation, we propose the variational word mask (VMASK) method to automatically learn task-specific important words and reduce irrelevant informati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
20
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 31 publications
(20 citation statements)
references
References 31 publications
(24 reference statements)
0
20
0
Order By: Relevance
“…For each example, LIME approximates the local decision boundary by fitting a linear model over the samples obtained by perturbing the example. To measure the faithfulness of the local explanations obtained using LIME, we measure the area over perturbation curve (AOPC) (Samek et al, 2017;Nguyen, 2018;Chen and Ji, 2020) which is defined as:…”
Section: Results On Interpretabilitymentioning
confidence: 99%
“…For each example, LIME approximates the local decision boundary by fitting a linear model over the samples obtained by perturbing the example. To measure the faithfulness of the local explanations obtained using LIME, we measure the area over perturbation curve (AOPC) (Samek et al, 2017;Nguyen, 2018;Chen and Ji, 2020) which is defined as:…”
Section: Results On Interpretabilitymentioning
confidence: 99%
“…Recently, there are applications and advances of local explanation methods [23,30,32]. For instance in NLP, some analyze the contributions of segments in documents to positive and negative sentiments [4,8,9,25]. Some move forwards to finding segments towards text similarity [10], retrieving a text span towards question-answering [27], and making local explanation as alignment model in machine translation [1].…”
Section: Related Work and Discussionmentioning
confidence: 99%
“…Traditionally, off-the-shelf local explanation frameworks, such as the Shapley value in game theory [32] and the learning-based Local Interpretable Model-agnostic Explanation (LIME) [30] have been shown to work well on classification tasks with a small number of classes. In particular, there has been work on image classification [30], sentiment analysis [8], and evidence selection for question answering [27]. However, to the best of our knowledge, there has been less work studying explanations over models with sequential output and large class sizes at each time step.…”
Section: Introductionmentioning
confidence: 99%
“…This paper takes a closer look into the gap between user need and current XAI. Specifically, we survey the common forms of explanations, such as feature attribution [6,26], decision rule [43,22], or probe [30,10], used in 218 recent NLP papers, and compare them to the 43 questions collected in the XAI Question Bank [28]. We use the forms of the explanations to gauge the misalignment between user questions and current NLP explanations.…”
Section: Introductionmentioning
confidence: 99%
“…Explainable AI Formats-I 1-Feature Attribution (FAT) [43.99%] : highlight the subsequences in input texts [6,26], Typical question [34]:…”
Section: Introductionmentioning
confidence: 99%