2018 Conference on Cognitive Computational Neuroscience 2018
DOI: 10.32470/ccn.2018.1234-0
|View full text |Cite
|
Sign up to set email alerts
|

Can Deep Neural Networks Rival Human Ability to Generalize in Core Object Recognition?

Abstract: Humans are thought to transfer their knowledge well to unseen domains. This putative ability to generalize is often juxtaposed against deep neural networks that are believed to be mostly domain-specific. Here we assessed the extent of generalization abilities in humans and ImageNet-trained models along two axes of image variations: perturbations to images (e.g., shuffling, blurring) and changes in representation style (e.g., paintings, cartoons). We found that models often matched or exceeded human performance… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
10
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
3
3

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(11 citation statements)
references
References 10 publications
1
10
0
Order By: Relevance
“…This is also the case with CNNs widely used in computer science that are often described as the best current theory of object recognition in humans (e.g. Kubilius, Kar, Schmidt, & DiCarlo, 2018 ). Although the design of CNNs (both the convolution and the pooling layers) are claimed to support translation invariance (for introduction see O'Shea & Nash, 2015 ), these models are generally trained to categorize images by training multiple exemplars of each image category at multiple spatial locations.…”
Section: Introductionmentioning
confidence: 93%
“…This is also the case with CNNs widely used in computer science that are often described as the best current theory of object recognition in humans (e.g. Kubilius, Kar, Schmidt, & DiCarlo, 2018 ). Although the design of CNNs (both the convolution and the pooling layers) are claimed to support translation invariance (for introduction see O'Shea & Nash, 2015 ), these models are generally trained to categorize images by training multiple exemplars of each image category at multiple spatial locations.…”
Section: Introductionmentioning
confidence: 93%
“…Presently, our datasets use naturalistic images, generated by pasting objects on a random backgrounds. While these datasets are already extremely challenging, we will more stringently be able to test model ability to generalize beyond its training set by expanding our datasets to more classes of images (e.g., photographs, distorted images (Geirhos et al, 2018), artistic renderings (Kubilius et al, 2018a), images optimized for neural responses ).…”
Section: By Acquiring the Same Types Of Data Using New Imagesmentioning
confidence: 99%
“…The quantitative relationship of this axis to the real-world size of objects was evaluated with a series of functions in different scales, 𝐿𝑖𝑛𝑒𝑎𝑟: 𝑦 = 𝑥 (6) 𝐸𝑥𝑝𝑜𝑛𝑒𝑛𝑡𝑖𝑎𝑙: 𝑦 = 𝑥 $. ** (7) 𝐸𝑥𝑝𝑜𝑛𝑒𝑛𝑡𝑖𝑎𝑙: 𝑦 = 𝑥 $.+ (8) 𝐸𝑥𝑝𝑜𝑛𝑒𝑛𝑡𝑖𝑎𝑙: 𝑦 = 𝑥 % (9) 𝐸𝑥𝑝𝑜𝑛𝑒𝑛𝑡𝑖𝑎𝑙: 𝑦 = 𝑥 * (10) 𝐿𝑜𝑔𝑎𝑟𝑖𝑡ℎ𝑚𝑖𝑐: 𝑦 = ln 𝑥 (11) where x is the value of the real-world size PC, and y is the measured real-world size of objects. The scale with the best fit was taken as the relationship that best describes the data.…”
Section: Evaluate the Role Of Real-world Size In Object Spacementioning
confidence: 99%
“…Recently, deep convolutional neural networks (DCNNs) show great potentials of simulating primates' ventral visual pathway on object recognition [9][10][11][12][13] , with similar representations between DCNNs and primates been revealed (e.g., retinotopy 14 , semantic structure 15 , coding scheme 16 , and face representation 17 ).…”
Section: Introductionmentioning
confidence: 99%