2022
DOI: 10.31234/osf.io/5zf4s
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Problems with Neural Network Models of Human Vision

Abstract: Deep neural networks (DNNs) have had extraordinary successes in classifying photographic images of objects and are often described as the best models of biological vision. This conclusion is largely based on three sets of findings: (1) DNNs are more accurate than any other model in classifying images taken from various datasets, (2) DNNs do the best job in predicting the pattern of human errors in classifying objects taken from various behavioral benchmark datasets, and (3) DNNs do the best job in predicting … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 23 publications
(15 citation statements)
references
References 102 publications
2
12
0
Order By: Relevance
“…We have argued for this approach in relation to making inferences about mechanistic similarity between DNNs and humans [29]. In fact, research relating DNNs to human vision provides a striking case of a disconnect between RSA and behavioural findings from psychology [2931]. The findings here may explain contradictory RSA scores between DNNs and human visual processing as pointed out by Xu and Vaziri-Pashkam [20].…”
Section: Discussionmentioning
confidence: 64%
See 1 more Smart Citation
“…We have argued for this approach in relation to making inferences about mechanistic similarity between DNNs and humans [29]. In fact, research relating DNNs to human vision provides a striking case of a disconnect between RSA and behavioural findings from psychology [2931]. The findings here may explain contradictory RSA scores between DNNs and human visual processing as pointed out by Xu and Vaziri-Pashkam [20].…”
Section: Discussionmentioning
confidence: 64%
“…However, another approach is more tractable: conduct controlled experiments to establish whether the two systems are representing information in similar ways. We have argued for this approach in relation to making inferences about mechanistic similarity between DNNs and humans [29]. In fact, research relating DNNs to human vision provides a striking case of a disconnect between RSA and behavioural findings from psychology [2931].…”
Section: Discussionmentioning
confidence: 99%
“…These failures may reflect a range of processes present in humans but absent in CNNs trained to recognise objects through supervised learning, such as figure-ground segregation, completing objects behind occluders, encoding border ownership, and inferring 3D properties about the object [43]. Consistent with this hypothesis, Jacob et al [27] and Bowers et al [7] have recently highlighted a number of these failures in CNNs, including a failure to represent 3D structure, occlusion, and parts of objects. More broadly, these results challenge the the claim that CNNs trained to recognise objects through supervised learning are good models of the ventral visual stream of human vision (see, for example, [52, 8, 37]).…”
Section: Discussionmentioning
confidence: 99%
“…But using ANNs in that way requires careful construction and comparison to ensure meaningful inferences can be drawn precisely because of these (and other) disanalogies baked into the technology. Failure to account for this can lead to misleading conclusions and faulty science [14]. On the other hand, nothing we have said means that ANNs cannot be brought further in line with biology to fruitful ends.…”
Section: Ontological Unificationmentioning
confidence: 97%
“…On a basic level, the tendency to ignore domain experts (Section 4.1) and the issues around path dependency (Section 4.2) may make ML systems less effective tools, which for systems that are so widely used has immediate social welfare implications. 14 And as we discussed in Section 4.4, the trend of increased black-boxing itself has distinct ethical implications related to subjects' rights to explanations.…”
Section: Ethicalmentioning
confidence: 99%