2020
DOI: 10.48550/arxiv.2011.09594
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Multi-view Depth Estimation with Predicted Uncertainty

Abstract: In this paper, we address the problem of estimating dense depth from a sequence of images using deep neural networks. Specifically, we employ a dense-optical-flow network to compute correspondences and then triangulate the point cloud to obtain an initial depth map. Parts of the point cloud, however, may be less accurate than others due to lack of common observations or small baseline-to-depth ratio.To further increase the triangulation accuracy, we introduce a depth-refinement network (DRN) that optimizes the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 32 publications
(58 reference statements)
0
2
0
Order By: Relevance
“…This predicted optical flow is used to generate a depth map via triangulation [16]. For every pixel x t i = [x t i , y t i , 1] T in frame t, its predicted optical flow corresponds to the displacement ôi between frames I t and I t+1 , such that x t+1 i = x t i + ôi .…”
Section: B Depth Triangulationmentioning
confidence: 99%
See 1 more Smart Citation
“…This predicted optical flow is used to generate a depth map via triangulation [16]. For every pixel x t i = [x t i , y t i , 1] T in frame t, its predicted optical flow corresponds to the displacement ôi between frames I t and I t+1 , such that x t+1 i = x t i + ôi .…”
Section: B Depth Triangulationmentioning
confidence: 99%
“…Using geometric view synthesis as a learning objective, this approach has been successfully applied to a wide range of key robot vision tasks in the challenging monocular setting, including the estimation of ego-motion [1], [2], the 6 degree-of-freedom camera translation and rotation; depth [3], [4], [5], [6], the perpixel distance value from the image plane; optical flow [7], [8], [9], [10], [11], the 2D pixel displacement between frames; and scene flow [12], [13], [14], [15], the 3D motion of each point in the scene. Although these tasks are clearly related [12], [13], [16], they typically require stereo pairs at training time [14], [15] to resolve reprojection ambiguities in self-supervision.…”
Section: Introductionmentioning
confidence: 99%