2019
DOI: 10.1038/s41467-019-08987-4
|View full text |Cite
|
Sign up to set email alerts
|

Unmasking Clever Hans predictors and assessing what machines really learn

Abstract: Current learning machines have successfully solved hard application problems, reaching high accuracy and displaying seemingly intelligent behavior. Here we apply recent techniques for explaining decisions of state-of-the-art learning machines and analyze various tasks from computer vision and arcade games. This showcases a spectrum of problem-solving behaviors ranging from naive and short-sighted, to well-informed and strategic. We observe that standard performance evaluation metrics can be oblivious to distin… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
661
1
4

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 808 publications
(677 citation statements)
references
References 111 publications
4
661
1
4
Order By: Relevance
“…Recently, the increasing popularity of explainable AI methods (see e.g. [141,142,143,144]) have allowed us to gain insight into the inner workings of deep learning algorithms. In this manner, it has become possible to extract how a problem is solved by the deep model.…”
Section: Explainable Aimentioning
confidence: 99%
“…Recently, the increasing popularity of explainable AI methods (see e.g. [141,142,143,144]) have allowed us to gain insight into the inner workings of deep learning algorithms. In this manner, it has become possible to extract how a problem is solved by the deep model.…”
Section: Explainable Aimentioning
confidence: 99%
“…erefore, it is important that users can understand when the system will fail. As detecting errors is a claimed utility of instance-level explanations [36,50], we suggest that future work should evaluate this empirically in more detail. Our study design did not allow to draw conclusions in this regard because we did not fully counterbalance the order of tasks and True Negatives (TN) were not part of the task set.…”
Section: E Utility Of Saliency Maps Exists But It Is Limitedmentioning
confidence: 99%
“…In contrast, CNNs look for pa erns in a sub-symbolic fashion that lead to an outcome [7,39]. Because CNNs do not process data in a semantic fashion, other pa erns in an image (which may not belong to the concept) can contribute towards a classi cation outcome in unexpected ways [36]. An implication for the design is that we need to develop explanation algorithms that bridge the gap between humans and machines by leading the user to understand that the system is not basing its classi cation decision on higher-level semantics of the image.…”
Section: Reasoning On Examplesmentioning
confidence: 99%
See 1 more Smart Citation
“…These performance gains and the resistance to the dimensionality curse are enabled by the hierarchical processing inherent in these multilayer deep networks, which is a biomimetic property common to biological cortical networks (Poggio et al, 2017). However, training these deep networks requires large amounts of labelled data and usually results in a black-box transformation, without any transparent internal mechanisms that would generate insights into the underlying control scheme (reviewed in Lapuschkin et al, 2019). In addition, machine learning solutions often require episodic model retraining (Hermann et al, 2015), and rely on a considerable memory space to store the necessary parameters (Weston et al, 2014).…”
Section: Introductionmentioning
confidence: 99%