2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.76
|View full text |Cite
|
Sign up to set email alerts
|

Split-Brain Autoencoders: Unsupervised Learning by Cross-Channel Prediction

Abstract: We propose split-brain autoencoders, a straightforward modification of the traditional autoencoder architecture, for unsupervised representation learning. The method adds a split to the network, resulting in two disjoint sub-networks. Each sub-network is trained to perform a difficult taskpredicting one subset of the data channels from another. Together, the sub-networks extract features from the entire input signal. By forcing the network to solve crosschannel prediction tasks, we induce a representation with… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
511
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 599 publications
(514 citation statements)
references
References 31 publications
2
511
1
Order By: Relevance
“…Here, we focused on the proxy task. Like most previous self-supervised work, we used a simple AlexNet architecture (17)(18)(19)29,30). However, future optimizations to the architecture will likely improve the applicability and performance of our method.…”
Section: Discussionmentioning
confidence: 99%
See 4 more Smart Citations
“…Here, we focused on the proxy task. Like most previous self-supervised work, we used a simple AlexNet architecture (17)(18)(19)29,30). However, future optimizations to the architecture will likely improve the applicability and performance of our method.…”
Section: Discussionmentioning
confidence: 99%
“…We show a summary of our architecture in Figure 1C. Following other work in self-supervised learning (17,18,30), we use an AlexNet architecture for the source cell encoder, although we set all kernel sizes to 3 due to the smaller sizes of our image patches, and we add batch normalization after each convolutional layer. We use a smaller number of filters and fewer convolutional layers in the architecture of the target marker encoder; we use three convolutional layers, with 16, 32, and 32 filters, respectively.…”
Section: Architecturementioning
confidence: 99%
See 3 more Smart Citations