2018 International Conference on 3D Vision (3DV) 2018
DOI: 10.1109/3dv.2018.00044
|View full text |Cite
|
Sign up to set email alerts
|

Structured Adversarial Training for Unsupervised Monocular Depth Estimation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 39 publications
(17 citation statements)
references
References 21 publications
0
16
0
Order By: Relevance
“…[16] showed that the inclusion of a local structure based appearance loss [65] significantly improved depth estimation performance compared to simple pairwise pixel differences [67,13,76]. [29] extended this approach to include an error fitting term, and [44] explored combining it with an adversarial based loss to encourage realistic looking synthesized images. Finally, inspired by [72], [73] use ground truth depth to train an appearance matching term.…”
Section: Appearance Based Lossesmentioning
confidence: 99%
“…[16] showed that the inclusion of a local structure based appearance loss [65] significantly improved depth estimation performance compared to simple pairwise pixel differences [67,13,76]. [29] extended this approach to include an error fitting term, and [44] explored combining it with an adversarial based loss to encourage realistic looking synthesized images. Finally, inspired by [72], [73] use ground truth depth to train an appearance matching term.…”
Section: Appearance Based Lossesmentioning
confidence: 99%
“…With Synthetic Depth: Synthetic data is an interesting source of ground-truth depths and/or stereo pairs. Instead of the usual photometric loss, domain adaptation is possible using generative adversarial networks [25], or by leveraging the ability of stereo matching networks to better generalize to real world data [11]. Luo et al [21] demonstrate how synthetic data can be incorporated into single-image depth estimation with a two stage process.…”
Section: Additional Supervisionmentioning
confidence: 99%
“…This loss is usually the distance between a reference image and the depth-guided reprojection of other views into that reference viewpoint. Depth regression is optimized and relative poses come from stereo camera calibration in a training-from-stereo setting [9,7,25,27,26], while depth values and camera poses can be optimized jointly when training on videos [42,20,37,22,38,32,44,36,28].…”
Section: Introductionmentioning
confidence: 99%
“…Departing from conventional CNN, new technologies are being introduced. Adversarial learning of Generative Adversarial Network (GAN) is applied to train MDE models [25], [26]. Depth and pose prediction models are used as a generator to synthesize a target image, and a discriminator distinguishes real images from synthesized images.…”
Section: Related Workmentioning
confidence: 99%