2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.01392
|View full text |Cite
|
Sign up to set email alerts
|

Patch-NetVLAD: Multi-Scale Fusion of Locally-Global Descriptors for Place Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
196
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 223 publications
(199 citation statements)
references
References 79 publications
1
196
0
Order By: Relevance
“…MSW performed better than object proposal techniques due to the use of the sliding window, especially when illumination or viewpoint changes. Similar to MSW, Patch-NetVLAD [10] also used the sliding window to generate multi-scale patches and obtained patch descriptors from NetVLAD residuals. However, these patches are described using NetVLAD trained on the whole image, which is not accurate.…”
Section: Descriptor-centered Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…MSW performed better than object proposal techniques due to the use of the sliding window, especially when illumination or viewpoint changes. Similar to MSW, Patch-NetVLAD [10] also used the sliding window to generate multi-scale patches and obtained patch descriptors from NetVLAD residuals. However, these patches are described using NetVLAD trained on the whole image, which is not accurate.…”
Section: Descriptor-centered Methodsmentioning
confidence: 99%
“…These global descriptors [1,6,7] which describe the whole image typically excel in terms of their robustness to appearance and illumination changes, as they are directly optimized for place recognition. However, as shown in [10,23,29], in areas with similar scenes, the global descriptor has difficulty distinguishing differences between local regions.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…One focus within VPR research is around designing methods that can successfully match places across changing appearance conditions such as day/night and seasonal changes [1], [4], [8], as well as across viewpoint changes [21], [25]. VPR approaches can be further partitioned into methods that perform matching using deep learning [4], [5], [8], [26] (see [27] for a review) or handcrafted visual features [1], [3] (see [28] for a review). Common across many of these approaches is a match scoring system; given a pair of images, a typical VPR system outputs a "quality score" which indicates the likelihood of two images being captured from the same place.…”
Section: A Visual Place Recognitionmentioning
confidence: 99%
“…More recently, NetVLAD [4] first proposed an end-to-end place recognition network by using a CNN for feature extraction and a differentiable NetVLAD layer for global aggregation. This inspired a series of end-to-end visual place recognition networks [5], [6], [7], [8], [9]. To utilise 3-D information, PointNetVLAD [10] and its successors [11], [12], [13], [14], [15] use point clouds as inputs and achieve very high average recall rates in outdoor environments with 3-D features aggregated Fig.…”
Section: Introductionmentioning
confidence: 99%