2019 International Conference on Robotics and Automation (ICRA) 2019
DOI: 10.1109/icra.2019.8794387
|View full text |Cite
|
Sign up to set email alerts
|

Night-to-Day Image Translation for Retrieval-based Localization

Abstract: Visual localization is a key step in many robotics pipelines, allowing the robot to (approximately) determine its position and orientation in the world. An efficient and scalable approach to visual localization is to use image retrieval techniques. These approaches identify the image most similar to a query photo in a database of geo-tagged images and approximate the query's pose via the pose of the retrieved database image. However, image retrieval across drastically different illumination conditions, e.g. da… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
130
0
3

Year Published

2019
2019
2020
2020

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 186 publications
(149 citation statements)
references
References 26 publications
(66 reference statements)
1
130
0
3
Order By: Relevance
“…A recent work [21] proposes a method to improve mining of triplets composing of hard negatives for training. A few works have addressed the problem of seasonal or day-night variations either by using 3D point clouds [16] or by domain transfer [13]. Others have proposed better or faster matching [22], [23], facilitating image retrieval.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…A recent work [21] proposes a method to improve mining of triplets composing of hard negatives for training. A few works have addressed the problem of seasonal or day-night variations either by using 3D point clouds [16] or by domain transfer [13]. Others have proposed better or faster matching [22], [23], facilitating image retrieval.…”
Section: Related Workmentioning
confidence: 99%
“…Moreover, it also means that precise localization may not be possible with such a feature embedding. Although several methods have shown accurate localization by learning features from densely distributed images [11], [12], [13], this may not always be feasible due to high computational and memory requirements. We argue that learning features whose distances are directly proportional to the geometric counterpart in the map, results in more versatile and powerful features which also provide higher retrieval accuracy.…”
Section: Introductionmentioning
confidence: 99%
“…Style-Transfer: Other approaches attempt to directly train computer vision models using synthetic data generated via style-transfer, or to directly adapt the input data to the target domain. Notable approaches include those of [4], [5], [23], [3] and [2]. Generally, these methods seem to have the most promise of reducing the domain gap between real and synthetic images, hence our decision to generate training data using the approach of [24].…”
Section: B Domain Adaptationmentioning
confidence: 99%
“…As a type of domain adaptation technique, domain unification is the holy grail of visual perception, theoretically allowing models trained on samples with limited heterogeneity to perform adequately on scenes that are well out of the distribution of the training data. Domain unification can be applied within the vast distribution of natural images [1], [2], [3], between natural and synthetic images (computer-generated, whether through traditional 3D rendering or more modern GAN-based techniques) [4], [5] and even between different sensor modalities [6]. Additionally, domain unification can be implemented at different stages of a computer vision pipeline, ranging from direct approaches such as domain confusion [7], [8], [9], fine-tuning models on target domains [1] or mixture-of-expert approaches [10], etc.…”
Section: Introductionmentioning
confidence: 99%
“…Lowry et al [2] proposed a simple approach based on using modified PCA to remove dimensions of variant conditions and showed impressive results. Adversia Porav et al [3] and Anoosheh et al [4] both overcame condition variance through image translation. Yin et al [5] proposed to separate condition-invariant features from extracted features using a CNN.…”
Section: Introductionmentioning
confidence: 99%