2015
DOI: 10.1101/032623
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Dynamics of scene representations in the human brain revealed by magnetoencephalography and deep neural networks

Abstract: Human scene recognition is a rapid multistep process evolving over time from single scene image to spatial layout processing. We used multivariate pattern analyses on magnetoencephalography (MEG) data to unravel the time course of this cortical process. Following an early signal for lowerlevel visual analysis of single scenes at ~100 ms, we found a marker of real-world scene size, i.e. spatial layout processing, at ~250 ms indexing neural representations robust to changes in unrelated scene properties and view… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

6
66
1

Year Published

2017
2017
2020
2020

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 52 publications
(73 citation statements)
references
References 62 publications
6
66
1
Order By: Relevance
“…previous work using MEG to correlate neural responses with computer vision features focused mainly on temporal patterns(Cichy, Khosla, Pantazis, & Oliva, 2017;Cichy, Khosla, Pantazis, Torralba, & Oliva, 2016c;Clarke et al, 2014). In line with the standard hierarchical model of primate visual cortex, these prior studies observed an earlyto-late shift from lower-level to higher-level features-a result consistent with the results we present here.…”
supporting
confidence: 91%
See 2 more Smart Citations
“…previous work using MEG to correlate neural responses with computer vision features focused mainly on temporal patterns(Cichy, Khosla, Pantazis, & Oliva, 2017;Cichy, Khosla, Pantazis, Torralba, & Oliva, 2016c;Clarke et al, 2014). In line with the standard hierarchical model of primate visual cortex, these prior studies observed an earlyto-late shift from lower-level to higher-level features-a result consistent with the results we present here.…”
supporting
confidence: 91%
“…However, in contrast to our present study, these studies focused on recognizing isolated objects on blank backgrounds (Cichy, Khosla, Pantazis, Torralba, & Oliva, 2016c;Clarke et al, 2014), or on specific properties of scenes, restricting the input to a small set of scene stimuli (Cichy, Khosla, Pantazis, & Oliva, 2017). However, in contrast to our present study, these studies focused on recognizing isolated objects on blank backgrounds (Cichy, Khosla, Pantazis, Torralba, & Oliva, 2016c;Clarke et al, 2014), or on specific properties of scenes, restricting the input to a small set of scene stimuli (Cichy, Khosla, Pantazis, & Oliva, 2017).…”
contrasting
confidence: 70%
See 1 more Smart Citation
“…Our results also shine new light on the temporal processing cascade during scene perception. Sensitivity to spatial structure emerged after 255 ms of processing, which is only after scene-selective peaks in ERPs (Harel et al, 2016;Sato et al, 1999) 9 and after basic scene attributes are computed (Cichy, Khosla, Pantazis, & Oliva, 2017).…”
Section: Discussionmentioning
confidence: 99%
“…A CNN is not designed to be biologically plausible, although we may draw a loose analogies between a CNN and the human cortex about their processing stages and representations (VanRullen, 2017). Comparative fMRI studies show that representations in the CNN correlate with emerging visual representations in the human brain (Cichy, Khosla, Pantazis, & Oliva, 2017;Cichy, Khosla, Pantazis, Torralba, & Oliva, 2016). Semantic similarity corresponds to more processed visual representations found in final CNN layers (in our case, fc7) or the inferotemporal cortex in the brain, whereas perceptual similarity is based on low-level representations in lower layers of a CNN or in the visual cortex.…”
Section: Discussionmentioning
confidence: 99%