2020
DOI: 10.1177/0278364920931151
|View full text |Cite
|
Sign up to set email alerts
|

Large-scale, real-time visual–inertial localization revisited

Abstract: The overarching goals in image-based localization are scale, robustness, and speed. In recent years, approaches based on local features and sparse 3D point-cloud models have both dominated the benchmarks and seen successful real-world deployment. They enable applications ranging from robot navigation, autonomous driving, virtual and augmented reality to device geo-localization. Recently, end-to-end learned localization approaches have been proposed which show promising results on small-scale datasets.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
68
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
1

Relationship

2
7

Authors

Journals

citations
Cited by 83 publications
(68 citation statements)
references
References 120 publications
0
68
0
Order By: Relevance
“…However, in many other applications, e.g., autonomous driving, the cameras record image sequences rather than taking individual pictures. There is a significant body of work on using sequences for visual localization [23,28,35,54,55,58,61,71,84,91]. These works combine single-image localization against a pre-built scene representation with local camera tracking based on visual and/or inertial odometry.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…However, in many other applications, e.g., autonomous driving, the cameras record image sequences rather than taking individual pictures. There is a significant body of work on using sequences for visual localization [23,28,35,54,55,58,61,71,84,91]. These works combine single-image localization against a pre-built scene representation with local camera tracking based on visual and/or inertial odometry.…”
Section: Introductionmentioning
confidence: 99%
“…Inspired by [28,51], this coarse estimate is used as a prior for a single-image localization module that provides pose estimates with respect to a global map. Finally, a fine localization stage similar to [35,54,55,58,61,84] refines the pose information provided by the single-image algorithm.…”
Section: Introductionmentioning
confidence: 99%
“…Liu et al (2018) and Dubé et al (2020) addressed 3D lidar-based SLAM methods on natural terrains and semi-unstructured environments, respectively, but not in collapsed, cluttered, and unstructured outdoor environments where assumptions such as segments and planes are not valid. Recent works on largescale and real-time visual SLAM show high-quality maps and loop-closing in outdoor urban environments (Lynen et al, 2020;Tanner et al, 2020), where GPS and 3D lidar are used as ground truth. Moreover, works in urban environments using vision, semantic mapping (Cadena et al, 2016), and including TIR images (Shin and Kim, 2019) offer promising techniques for potential SLAM applications in disaster robotics.…”
Section: Discussionmentioning
confidence: 99%
“…In the training phase, the calculative scale is very large, and GPU servers are usually required, which cannot run smoothly on mobile platforms [20]. In many scenarios, learning-based features are not as effective as traditional features such as SIFT, and the interpretability is poor [24][25][26][27].…”
Section: Related Workmentioning
confidence: 99%