2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.695
|View full text |Cite
|
Sign up to set email alerts
|

CNN-SLAM: Real-Time Dense Monocular SLAM with Learned Depth Prediction

Abstract: Given the recent advances in depth prediction from Convolutional Neural Networks (CNNs), this paper investigates how predicted depth maps from a deep neural network can be deployed for accurate and dense monocular reconstruction. We propose a method where CNN-predicted dense depth maps are naturally fused together with depth measurements obtained from direct monocular SLAM. Our fusion scheme privileges depth prediction in image locations where monocular SLAM approaches tend to fail, e.g. along low-textured reg… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
407
0
7

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 656 publications
(448 citation statements)
references
References 28 publications
0
407
0
7
Order By: Relevance
“…The photometric factor is treated as a baseline system and the other two types are included both individually and together. We perform this evaluation on selected shortened scenes from the validation set of the ScanNet dataset and use three error metrics: RMSE of the Absolute Trajectory Error (ATE-RMSE), the absolute relative depth difference (absrel) [13] and the average percentage of pixels in which the estimated depth falls within 10% of the true value (pc110) [18]. The results are presented in Table I.…”
Section: B Ablation Studiesmentioning
confidence: 99%
See 2 more Smart Citations
“…The photometric factor is treated as a baseline system and the other two types are included both individually and together. We perform this evaluation on selected shortened scenes from the validation set of the ScanNet dataset and use three error metrics: RMSE of the Absolute Trajectory Error (ATE-RMSE), the absolute relative depth difference (absrel) [13] and the average percentage of pixels in which the estimated depth falls within 10% of the true value (pc110) [18]. The results are presented in Table I.…”
Section: B Ablation Studiesmentioning
confidence: 99%
“…Ours CNN-SLAM [18] LSD-BS [19] Laina [37] icl/office0 [24]. CNN-SLAM was run without pose graph optimisation and our system without loop closures.…”
Section: Sequencementioning
confidence: 99%
See 1 more Smart Citation
“…Combined with SLAM systems, 2D semantic segmentation can be achieved in 3D environments [RA17] [TTLN17] [ZSS17] [MHDL17], a promising future in robotic vision understanding and autonomous driving. Unlike these existing methods that aimed at providing the semantic understanding of the scene for robots, we are focusing our attention on human interactions.…”
Section: Previous Workmentioning
confidence: 99%
“…Simultaneous Localization and Mapping (SLAM) is the process of building the map of an unknown environment and determining the location of a robot concurrently using this map. Recently introduced SLAM algorithms have objectoriented design [63], [64], [65], [66], [67], [68], [69], [70], [71]. Accurate object pose parameters provide camera-object constraints and lead to better performance on localization and mapping.…”
Section: Introductionmentioning
confidence: 99%