2020
DOI: 10.1101/2020.09.06.284794
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Self-Supervised Natural Image Reconstruction and Large-Scale Semantic Classification from Brain Activity

Abstract: Reconstructing natural images and decoding their semantic category from fMRI brain recordings is challenging. Acquiring sufficient pairs (image, fMRI) that span the huge space of natural images is prohibitive. We present a novel self-supervised approach for fMRI-to-image reconstruction and classification that goes well beyond the scarce paired data. By imposing cycle consistency, we train our image reconstruction deep neural network on many “unpaired” data: a plethora of natural images without fMRI recordings … Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(11 citation statements)
references
References 65 publications
0
11
0
Order By: Relevance
“…The use of CNNs is ubiquitous in image processing tasks, including image reconstruction. Specifically, encoder-decoder (Beliy et al, 2019;Gaziv et al, 2020), U-Net (Fang et al, 2020), generative adversarial network (Goodfellow et al, 2014), and variational autoencoder (Kingma and Welling, 2014) are popular architectures that adopt stacked convolutional layers to extract features at multiple levels. Shen et al (2019b) utilized a pretrained VGG-19-based DNN to extract hierarchical features from stimuli images (see Figure 3A).…”
Section: Convolutional Neural Network (Cnn)mentioning
confidence: 99%
See 4 more Smart Citations
“…The use of CNNs is ubiquitous in image processing tasks, including image reconstruction. Specifically, encoder-decoder (Beliy et al, 2019;Gaziv et al, 2020), U-Net (Fang et al, 2020), generative adversarial network (Goodfellow et al, 2014), and variational autoencoder (Kingma and Welling, 2014) are popular architectures that adopt stacked convolutional layers to extract features at multiple levels. Shen et al (2019b) utilized a pretrained VGG-19-based DNN to extract hierarchical features from stimuli images (see Figure 3A).…”
Section: Convolutional Neural Network (Cnn)mentioning
confidence: 99%
“…In a follow-up study, Gaziv et al (2020) improved the reconstruction accuracy of BeliyEncDec by introducing a loss function based on the perceptual similarity measure (Zhang et al, 2018). To calculate perceptual similarity loss, the authors first extracted multilayer features from original and reconstructed images using VGG and then compared the extracted features layerwise.…”
Section: Deterministic Encoder-decoder Modelsmentioning
confidence: 99%
See 3 more Smart Citations