2019 IEEE/CVF International Conference on Computer Vision (ICCV) 2019
DOI: 10.1109/iccv.2019.00759
|View full text |Cite
|
Sign up to set email alerts
|

Spatial Correspondence With Generative Adversarial Network: Learning Depth From Monocular Videos

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 24 publications
(13 citation statements)
references
References 37 publications
0
13
0
Order By: Relevance
“…Their network learns from the ground truth labels in synthetic domain as well as the epipolar geometry of the real domain, thereby achieving competitive results. Wu et al [105] improve the architecture of the generator by utilizing a spatial correspondence module for feature matching and an attention mechanism for feature re-weighting.…”
Section: Semi-supervised Monocular Depth Estimationmentioning
confidence: 99%
“…Their network learns from the ground truth labels in synthetic domain as well as the epipolar geometry of the real domain, thereby achieving competitive results. Wu et al [105] improve the architecture of the generator by utilizing a spatial correspondence module for feature matching and an attention mechanism for feature re-weighting.…”
Section: Semi-supervised Monocular Depth Estimationmentioning
confidence: 99%
“…Other work concentrates on multi-view stereo (MVS), which operates on unordered image sets [45,40,42,33,37,14,63,62,61]. Not requiring the ground truth depth and camera poses during training, self-supervised MVS methods [42,33,37,14,63,62,61] leverage cost volumes to process sequences of frames at test time.…”
Section: Monocular Depth Estimationmentioning
confidence: 99%
“…Other work concentrates on multi-view stereo (MVS), which operates on unordered image sets [45,40,42,33,37,14,63,62,61]. Not requiring the ground truth depth and camera poses during training, self-supervised MVS methods [42,33,37,14,63,62,61] leverage cost volumes to process sequences of frames at test time. Compared with the base method of MVS, these methods can predict the depth using images captured by moving cameras and do not need camera poses during training time.…”
Section: Monocular Depth Estimationmentioning
confidence: 99%
“…p interp (r) = max MonoDis [9] AM3D [26] GS3D [10] ROI-10D [6] 3D [41], [42], it is significantly more difficult for approaches that are solely built upon monocular RGB input as predicting depth from monocular images is a challenging task [43], [44]. Furthermore, Average Orientation Similarity (AOS) is used to assess the ability of 3D-GCK to correctly predict the object orientation [32].…”
Section: Metricsmentioning
confidence: 99%