2015 IEEE International Conference on Computer Vision (ICCV) 2015
DOI: 10.1109/iccv.2015.465
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Learning of Spatiotemporally Coherent Metrics

Abstract: Current state-of-the-art classification and detection algorithms train deep convolutional networks using labeled data. In this work we study unsupervised feature learning with convolutional networks in the context of temporally coherent unlabeled data. We focus on feature learning from unlabeled video data, using the assumption that adjacent video frames contain semantically similar information. This assumption is exploited to train a convolutional pooling auto-encoder regularized by slowness and sparsity prio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
113
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 115 publications
(113 citation statements)
references
References 17 publications
0
113
0
Order By: Relevance
“…When P = 0 the network is an invariance-free sparse autoencoder with no additional temporal cost. When P = N , temporal coherence is encoded in every hidden layer neuron, similar to the architecture described in [6]. With values inbetween, a mixed representation of invariant and variable features forms.…”
Section: Cost Functionmentioning
confidence: 99%
See 2 more Smart Citations
“…When P = 0 the network is an invariance-free sparse autoencoder with no additional temporal cost. When P = N , temporal coherence is encoded in every hidden layer neuron, similar to the architecture described in [6]. With values inbetween, a mixed representation of invariant and variable features forms.…”
Section: Cost Functionmentioning
confidence: 99%
“…This method has improved object identity performance on the benchmark COIL100 dataset [11], applying the temporal coherence to the output of a Convolutional Neural Network [12], alongside sparsity in a deep invariant architecture [13] and during pre-training and network output regularization [14]. An architecture relatively similar in nature to ours is that used by Goroshin et al [6]. Temporal coherence is applied to sparse autoencoders, as part of a convolutional architecture.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…al. [3] Wang and Gupta [31]. Their approach was based on metric learning of related frames in video, but their approaches were not capable of learning long term dependencies since they assumed only a simple zeromean Gaussian innovation between frames.…”
Section: Related Workmentioning
confidence: 99%
“…A machine learning application of this paradigm is to use temporal coherence as a proxy for learning sensory representations without strong supervision or explicit labels [3][4] [5]. One approach is a Bayesian formulation in which we assume the learning system builds an internal model of the world p(x t ; θ) for explaining input streams x t using a system parameterized by θ. Predictive Coding proposes to adapt this model to reduce discrepancies between predictionsp(x t |x t−1 ; θ) and observations p(x t ).…”
Section: Introductionmentioning
confidence: 99%