2016 IEEE International Conference on Image Processing (ICIP) 2016
DOI: 10.1109/icip.2016.7532634
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised convolutional neural networks for motion estimation

Abstract: Traditional methods for motion estimation estimate the motion field F between a pair of images as the one that minimizes a predesigned cost function. In this paper, we propose a direct method and train a Convolutional Neural Network (CNN) that when, at test time, is given a pair of images as input it produces a dense motion field F at its output layer. In the absence of large datasets with ground truth motion that would allow classical supervised training, we propose to train the network in an unsupervised man… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
71
0

Year Published

2018
2018
2019
2019

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 87 publications
(71 citation statements)
references
References 20 publications
0
71
0
Order By: Relevance
“…These algorithms take a pair of images as input, and use a convolutional neural network to learn image features that capture the concept of optical flow from data. Several of these works require supervision in the form of ground truth flow fields [52], [53], [55], [56], while we build on a few that use an unsupervised objective [51], [54]. The spatial transform layer enables neural networks to perform both global parametric 2D image alignment [42] and dense spatial transformations [54], [57], [58] without requiring supervised labels.…”
Section: D Image Alignmentmentioning
confidence: 99%
“…These algorithms take a pair of images as input, and use a convolutional neural network to learn image features that capture the concept of optical flow from data. Several of these works require supervision in the form of ground truth flow fields [52], [53], [55], [56], while we build on a few that use an unsupervised objective [51], [54]. The spatial transform layer enables neural networks to perform both global parametric 2D image alignment [42] and dense spatial transformations [54], [57], [58] without requiring supervised labels.…”
Section: D Image Alignmentmentioning
confidence: 99%
“…They are based on an autoencoder design, allow for supervised end-to-end training, and enable fast inference during testing time. To alleviate the need for training data with ground truth in a specific domain, unsupervised [1,40,47,58,63,66] and semisupervised [31,65] alternatives have also been developed.…”
Section: Related Workmentioning
confidence: 99%
“…We used FlowNetC trained by data scheduling without fine-tuning as a baseline in the evaluation. To obtain an unbiased evaluation result, we trained and tested each of these networks on both Flying Chairs and Sintel dataset [Butler et al, 2012] three times. The average EPE is reported in Table 3.…”
Section: Methodsmentioning
confidence: 99%
“…We evaluatethe our performance of the proposed approach on three standard optical flow benchmarks: Flying Chairs [Dosovitskiy et al, 2015], Sintel [Butler et al, 2012], and KITTI [Geiger et al, 2012]. We compare the performance of the proposed approach to both supervised methods such as: FlowNet(S/C) [ [Ilg et al, 2017] uses several FlowNets and contains cascade training of the FlowNets in different phases.…”
Section: Benchmarkmentioning
confidence: 99%
See 1 more Smart Citation