2018
DOI: 10.1109/lra.2018.2869640
|View full text |Cite
|
Sign up to set email alerts
|

VLocNet++: Deep Multitask Learning for Semantic Visual Localization and Odometry

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
149
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 233 publications
(156 citation statements)
references
References 3 publications
1
149
0
Order By: Relevance
“…Neural networks, such as PoseNet (Kendall, Grimes, & Cipolla, 2015), VLocNet++ (Radwan, Valada, & Burgard, 2018) Although much progress has been reported in the area of deep learning-based localization, VO techniques are still dominated by classical keypoints matching algorithms, combined with acceleration data provided by inertial sensors. This is mainly due to the fact that keypoints detectors are computational efficient and can be easily deployed on embedded devices.…”
Section: Localizationmentioning
confidence: 99%
See 1 more Smart Citation
“…Neural networks, such as PoseNet (Kendall, Grimes, & Cipolla, 2015), VLocNet++ (Radwan, Valada, & Burgard, 2018) Although much progress has been reported in the area of deep learning-based localization, VO techniques are still dominated by classical keypoints matching algorithms, combined with acceleration data provided by inertial sensors. This is mainly due to the fact that keypoints detectors are computational efficient and can be easily deployed on embedded devices.…”
Section: Localizationmentioning
confidence: 99%
“…, or the approaches introduced inWalch et al (2017),Melekhov, Ylioinas, Kannala, and Rahtu (2017),Laskar, Melekhov, Kalia, and Kannala (2017),Brachmann and Rother (2018), or Sarlin, Debraine, Dymczyk, Siegwart, andCadena (2018), are using image data to estimate the 3D pose of a camera in an End2End fashion. Scene semantics can be derived together with the estimated pose(Radwan et al, 2018).LiDAR intensity maps are also suited for learning a real-time, calibration-agnostic localization for autonomous cars (Barsan, Wang, Pokrovsky, & Urtasun, 2018). The method uses a deep neural network to build a learned representation of the driving scene from LiDAR sweeps and intensity maps.…”
mentioning
confidence: 99%
“…Starting from this work, additional improvements have been proposed by introducing new geometric loss functions [15], by exploiting the uncertainty estimation of Bayesian CNNs [16], by including a data augmentation scheme based on synthetic depth information [17], or using the relative pose between two observations in a CNNs pipeline [18]. One of the many works that follow the idea presented in PoseNet is VLocNet++ [19].…”
Section: A Camera-only Approachesmentioning
confidence: 99%
“…However, unlike previous work focused on improving the retrieval [1,4,20,26,34,36] or matching [56,72], we focus on the pose verification stage, i.e., the problem of selecting the "best" pose from the n estimated poses. An alternative to the localization approaches outlined above is to train a CNN that directly regresses the camera pose from a given input image [9,12,32,33,50,78]. However, it was recently shown that such methods do not consistently outperform a simple image retrieval baseline [59].…”
Section: Related Workmentioning
confidence: 99%