2021
DOI: 10.48550/arxiv.2101.06930
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generative Counterfactuals for Neural Networks via Attribute-Informed Perturbation

Abstract: With the wide use of deep neural networks (DNN), model interpretability has become a critical concern, since explainable decisions are preferred in high-stake scenarios. Current interpretation techniques mainly focus on the feature attribution perspective, which are limited in indicating why and how particular explanations are related to the prediction. To this end, an intriguing class of explanations, named counterfactuals, has been developed to further explore the "what-if" circumstances for interpretation, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
4

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 29 publications
(26 reference statements)
0
4
0
Order By: Relevance
“…The first family of methods conditions the generative model on attributes, by e.g. using a conditional GAN [26,33,50,59,60]. This dependency on attribute information can restrict the applicability of these methods in scenarios where annotations are scarce.…”
Section: Related Workmentioning
confidence: 99%
“…The first family of methods conditions the generative model on attributes, by e.g. using a conditional GAN [26,33,50,59,60]. This dependency on attribute information can restrict the applicability of these methods in scenarios where annotations are scarce.…”
Section: Related Workmentioning
confidence: 99%
“…This is not a popular feature in the literature, likely because it is not an easy task to produce appropriate comparisons. Counterfactual approaches (such as [14,34,28]) are a particular case where the contrast is based on synthetic situations (values of features/attributes); another example is that of balanced explanations, which appear in [10].…”
Section: A Survey Of Relevant Xai Approaches In the Literaturementioning
confidence: 99%
“…These techniques are very related to the previously discussed perturbationbased mechanisms, but they come with a stronger guarantee, namely that the perturbations are guaranteed to change the model's prediction. The generation of counterfactual explanations has received significant attention in the NLP community [22,23,36,45,46]. In the simplest case, these counterfactuals can be generated by deleting words from the input text [23] or via rewrite-rules such as adding negations or shuffling words [45].…”
Section: Related Workmentioning
confidence: 99%