2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00987
|View full text |Cite
|
Sign up to set email alerts
|

VITAMIN-E: VIsual Tracking and MappINg With Extremely Dense Feature Points

Abstract: In this paper, we propose a novel indirect monocular SLAM algorithm called "VITAMIN-E," which is highly accurate and robust as a result of tracking extremely dense feature points. Typical indirect methods have difficulty in reconstructing dense geometry because of their careful feature point selection for accurate matching. Unlike conventional methods, the proposed method processes an enormous number of feature points by tracking the local extrema of curvature informed by dominant flow estimation. Because this… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
14
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 32 publications
(16 citation statements)
references
References 34 publications
(30 reference statements)
0
14
0
Order By: Relevance
“…Recent work investigates CPU-based approaches in combination with RGB-D sensing (e.g., Wald et al, 2018), PanopticFusion (Narita et al, 2019), and Voxblox++ (Grinvald et al, 2019). A sparser set of contributions addresses other sensing modalities, including monocular cameras (e.g., CNN-SLAM (Tateno et al, 2017), VSO (Lianos et al, 2018), VITAMIN-E (Yokozuka et al, 2019), and XIVO (Dong et al, 2017)) and lidar (Behley et al, 2019; Dubé et al, 2018).…”
Section: Related Workmentioning
confidence: 99%
“…Recent work investigates CPU-based approaches in combination with RGB-D sensing (e.g., Wald et al, 2018), PanopticFusion (Narita et al, 2019), and Voxblox++ (Grinvald et al, 2019). A sparser set of contributions addresses other sensing modalities, including monocular cameras (e.g., CNN-SLAM (Tateno et al, 2017), VSO (Lianos et al, 2018), VITAMIN-E (Yokozuka et al, 2019), and XIVO (Dong et al, 2017)) and lidar (Behley et al, 2019; Dubé et al, 2018).…”
Section: Related Workmentioning
confidence: 99%
“…Approaches in SLAM can generally be classified based on the type of sensors used to sense the environment. LiDAR SLAM uses LiDAR sensors at its core while visual SLAM uses cameras as a main sensor such as monocular-camera [2]- [6], stereo [7]- [9], RGB-D [8], [10], [11], etc. Monocular-based approaches, although widely adopted due to their simple and economical setup, suffer from recovering the metric scale, and that integrating with an IMU (either loosely-coupled or tightly-coupled), referred to with a prefix visual-inertial, offers a main advantage of solving scale ambiguity and providing a more robust navigation.…”
Section: Related Workmentioning
confidence: 99%
“…AutoStitch 5 , OpenPano 6 , issues with memory management [65], we could only be able to compare our system against Agisoft Photoscan 7 (AS) and Context Capture 8 (CC). We use the same dataset as that used in Pose estimation evaluation, and the computational results are presented in Tab.…”
Section: Pose Estimation Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…Bundle Adjustment [112], consisting on the joint optimization of a set of camera poses and points, is frequently used to obtain a globally consistent model of the scene [79]. However, there are also several recent VSLAM approaches that alternate the optimization between points and poses, reducing the computational cost with a small impact in the accuracy, given a sufficient number of points [84,94,120,123]. In its most basic form, the map model consists of a set of n points and m RGB-D keyframes.…”
Section: Point-based Mappingmentioning
confidence: 99%