2023
DOI: 10.3389/fncom.2023.1172883
|View full text |Cite
|
Sign up to set email alerts
|

Interpreting the decisions of CNNs via influence functions

Abstract: An understanding of deep neural network decisions is based on the interpretability of model, which provides explanations that are understandable to human beings and helps avoid biases in model predictions. This study investigates and interprets the model output based on images from the training dataset, i.e., to debug the results of a network model in relation to the training dataset. Our objective was to understand the behavior (specifically, class prediction) of deep learning models through the analysis of p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 45 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?