2017
DOI: 10.48550/arxiv.1711.06104
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Towards better understanding of gradient-based attribution methods for Deep Neural Networks

Abstract: Understanding the flow of information in Deep Neural Networks (DNNs) is a challenging problem that has gain increasing attention over the last few years. While several methods have been proposed to explain network predictions, there have been only a few attempts to compare them from a theoretical perspective. What is more, no exhaustive empirical comparison has been performed in the past. In this work, we analyze four gradient-based attribution methods and formally prove conditions of equivalence and approxima… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
165
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 114 publications
(166 citation statements)
references
References 15 publications
(23 reference statements)
1
165
0
Order By: Relevance
“…Similarly, Integrated Gradients computes the average gradient of the output with respect to each input feature by integrating from a baseline to the current feature value [35]. DeepLIFT (Deep Learning Important FeaTures) [36] works with deep NNs and it is a good approximation to Integrated Gradients [37]. Similar to integrated gradients it also defines a "reference activation" which is often viewed as "uninformative" in context, e.g.…”
Section: Feature Importance Methodsmentioning
confidence: 99%
“…Similarly, Integrated Gradients computes the average gradient of the output with respect to each input feature by integrating from a baseline to the current feature value [35]. DeepLIFT (Deep Learning Important FeaTures) [36] works with deep NNs and it is a good approximation to Integrated Gradients [37]. Similar to integrated gradients it also defines a "reference activation" which is often viewed as "uninformative" in context, e.g.…”
Section: Feature Importance Methodsmentioning
confidence: 99%
“…the feature of interest is computed pointwise after a smoothing procedure. Symbolic derivatives are commonly used to determine the importance of features for neural networks (Ancona et al 2017). While MEs provide interpretations in terms of prediction changes, most methods provide an interpretation in terms of prediction levels.…”
Section: Interpretable Machine Learningmentioning
confidence: 99%
“…where changes in activations would change the output most (the gradients). An overview of various methods for salience mapping is available elsewhere [32]. The class activation map (CAM)/grad-CAM [33,34] approach builds a map of the input regions that are responsible for a classification by calculating how the different convolutional filters contribute to that classification and building a weighted average of these activations, which can then be projected onto the input image, the operation of CAM is presented schematically in Figure 3.…”
Section: Deep Interpretationsmentioning
confidence: 99%