2018
DOI: 10.1523/jneurosci.0388-18.2018
|View full text |Cite
|
Sign up to set email alerts
|

Large-Scale, High-Resolution Comparison of the Core Visual Object Recognition Behavior of Humans, Monkeys, and State-of-the-Art Deep Artificial Neural Networks

Abstract: Primates, including humans, can typically recognize objects in visual images at a glance despite naturally occurring identity-preserving image transformations (e.g., changes in viewpoint). A primary neuroscience goal is to uncover neuron-level mechanistic models that quantitatively explain this behavior by predicting primate performance for each and every image. Here, we applied this stringent behavioral prediction test to the leading mechanistic models of primate vision (specifically, deep, convolutional, art… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

12
173
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 269 publications
(196 citation statements)
references
References 34 publications
(33 reference statements)
12
173
1
Order By: Relevance
“…Most intriguingly, this tuning tended to be monotonic rather than bell-shaped: The preferred stimulus for a neuron tended to be at the extreme end of the deformation range. This is quite different from prototype models of population coding for shape that assume that each neuron will have bell-shaped tuning around a preferred stimulus and that these preferred stimuli will be broadly distributed across the parameter space (e.g., Riesenhuber & Poggio 2002). Such bell-shaped tuning for 3D pose has been observed in IT for 3D objects (Logothetis et al 1995); however, variation in 3D pose results in discontinuous configural changes in the object projection (i.e., the occlusion or dis-occlusion of visual features) that form a natural partitioning of the view sphere into prototypes.…”
Section: High-level Neural Representations: Inferior Temporal Cortexmentioning
confidence: 88%
“…Most intriguingly, this tuning tended to be monotonic rather than bell-shaped: The preferred stimulus for a neuron tended to be at the extreme end of the deformation range. This is quite different from prototype models of population coding for shape that assume that each neuron will have bell-shaped tuning around a preferred stimulus and that these preferred stimuli will be broadly distributed across the parameter space (e.g., Riesenhuber & Poggio 2002). Such bell-shaped tuning for 3D pose has been observed in IT for 3D objects (Logothetis et al 1995); however, variation in 3D pose results in discontinuous configural changes in the object projection (i.e., the occlusion or dis-occlusion of visual features) that form a natural partitioning of the view sphere into prototypes.…”
Section: High-level Neural Representations: Inferior Temporal Cortexmentioning
confidence: 88%
“…Analogous to the brain's sensory network, DNNs perform complex computations through deep stacks of simple intra-layer neural circuits. Thus, researchers have widely used DNN models to understand the human brain network, especially sensory brain networks (Eickenberg, Gramfort, Varoquaux, & Thirion, 2017;Guclu & van Gerven, 2015;Horikawa & Kamitani, 2017;Rajalingham et al, 2018;Yamins & DiCarlo, 2016). At the same time, DNNs are capable of discovering complex structures within high-dimensional input data, and can transform these structures into abstract levels (LeCun et al, 2015).…”
Section: Deep Learning As a Research Toolmentioning
confidence: 99%
“…In addition to providing experimental evidence of well-studied principles that underlie human grasping, this data-driven exploration could improve our understanding of the function of touch during object manipulation. Deep-learning models have greatly advanced our knowledge of the neural mechanisms that underlie visual object recognition 3 . In this respect, a similar approach could be applied to the interpretation of tactile-information processing in the brain.…”
Section: A Step Forward For Artificial Touch Systemsmentioning
confidence: 99%