2017
DOI: 10.1016/j.patcog.2016.11.008
|View full text |Cite
|
Sign up to set email alerts
|

Explaining nonlinear classification decisions with deep Taylor decomposition

Abstract: Nonlinear methods such as Deep Neural Networks (DNNs) are the gold standard for various challenging machine learning problems such as image recognition. Although these methods perform impressively well, they have a significant disadvantage, the lack of transparency, limiting the interpretability of the solution and thus the scope of application in practice. Especially DNNs act as black boxes due to their multilayer nonlinear structure. In this paper we introduce a novel methodology for interpreting generic mul… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
957
0
1

Year Published

2017
2017
2023
2023

Publication Types

Select...
6
3

Relationship

2
7

Authors

Journals

citations
Cited by 1,150 publications
(1,006 citation statements)
references
References 34 publications
3
957
0
1
Order By: Relevance
“…Another future direction would be to analyze the interpretability of NNS systems, specifically for recommender systems with non-linear query mechanism, in terms of salient features that have led to the query result. This is in the line of research on ''explaining learning machines'', i.e., answering to the question which part of the data is responsible for specific decisions made by learning machines (Baehrens et al 2010;Zeiler and Fergus 2014;Bach et al 2015; Ribeiro et al 2016; Montavon et al 2017Montavon et al , 2018. This question is non-trivial when the learning machines are complex and non-linear.…”
Section: Resultsmentioning
confidence: 97%
“…Another future direction would be to analyze the interpretability of NNS systems, specifically for recommender systems with non-linear query mechanism, in terms of salient features that have led to the query result. This is in the line of research on ''explaining learning machines'', i.e., answering to the question which part of the data is responsible for specific decisions made by learning machines (Baehrens et al 2010;Zeiler and Fergus 2014;Bach et al 2015; Ribeiro et al 2016; Montavon et al 2017Montavon et al , 2018. This question is non-trivial when the learning machines are complex and non-linear.…”
Section: Resultsmentioning
confidence: 97%
“…This redistribution rule has been showed to fulfill the layer-wise conservation property [10] and to be closely related to a deep variant of Taylor decomposition [11].…”
Section: B Interpretabilitymentioning
confidence: 93%
“…LRP explains individual classification decisions of a DNN by decomposing its output in terms of input variables. It is a principled method which has close relation to Taylor decomposition [11] and is applicable to arbitrary DNN architectures. From a practitioners perspective LRP adds a new dimension to the application of DNNs (e.g., in computer vision [12], [13]) by making the prediction transparent.…”
Section: Introductionmentioning
confidence: 99%
“…It is a principled method that has a close relationship to Taylor decomposition and is applicable to arbitrary deep neural network architectures [30]. The output is a heatmap over the input features that indicates the relevance of each feature to the model output.…”
Section: B Model Functionalitymentioning
confidence: 99%