2023
DOI: 10.1109/tmm.2022.3155927
|View full text |Cite
|
Sign up to set email alerts
|

ALIKE: Accurate and Lightweight Keypoint Detection and Descriptor Extraction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
32
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 38 publications
(32 citation statements)
references
References 40 publications
0
32
0
Order By: Relevance
“…Concerning deep image matching methods, we included the current state-of-the-art according to recent evaluation (Remondino et al, 2021;Chen et al, 2021;Jin et al, 2021;Ma et al, 2021;Bellavia et al, 2022b): the DIScrete Keypoints (DISK, Tyszkiewicz et al, 2020), the Hybrid Pipeline (HP, Bellavia et al, 2022c) also without rotational invariance provided by OriNet (Mishkin et al, 2018) (denoted as HP_upright ), the Local Feature TRansformer (LoFTR, Sun et al, 2021) and its rotation invariant extension SE2-LoFTR (Bökman et al, 2022), SuperPoint+SuperGlue (Sarlin et al, 2020), the Accurate Shape and Localization Features (ASLFeat, Luo et al, 2020), the Repeatable Detector and Descriptor (R2D2, Revaud et al, 2019) and the Local Feature Network (LF-Net, Ono et al, 2018). Two further deep-learning methods providing quite interesting results were also added in this evaluation: the Rotation-Robust Descriptors (RoRD, Parihar et al, 2021) for its rotation invariance and the Accurate and Lightweight Keypoint Detection and Descriptor Extraction (ALIKE, Zhao et al, 2022) for its ability to run in real-time.…”
Section: Compared Methodsmentioning
confidence: 99%
“…Concerning deep image matching methods, we included the current state-of-the-art according to recent evaluation (Remondino et al, 2021;Chen et al, 2021;Jin et al, 2021;Ma et al, 2021;Bellavia et al, 2022b): the DIScrete Keypoints (DISK, Tyszkiewicz et al, 2020), the Hybrid Pipeline (HP, Bellavia et al, 2022c) also without rotational invariance provided by OriNet (Mishkin et al, 2018) (denoted as HP_upright ), the Local Feature TRansformer (LoFTR, Sun et al, 2021) and its rotation invariant extension SE2-LoFTR (Bökman et al, 2022), SuperPoint+SuperGlue (Sarlin et al, 2020), the Accurate Shape and Localization Features (ASLFeat, Luo et al, 2020), the Repeatable Detector and Descriptor (R2D2, Revaud et al, 2019) and the Local Feature Network (LF-Net, Ono et al, 2018). Two further deep-learning methods providing quite interesting results were also added in this evaluation: the Rotation-Robust Descriptors (RoRD, Parihar et al, 2021) for its rotation invariance and the Accurate and Lightweight Keypoint Detection and Descriptor Extraction (ALIKE, Zhao et al, 2022) for its ability to run in real-time.…”
Section: Compared Methodsmentioning
confidence: 99%
“…DISK [22] uses reinforcement learning to train the score map and descriptor map. ALIKE [10] has a differentiable keypoint detection module for accurate keypoint training and has the lightest network, thereby allowing its application in real-time visual measurement applications. D2-Net [23] does not estimate the score map with the network, but rather detects keypoints with II).…”
Section: B Joint Keypoint and Descriptor Learningmentioning
confidence: 99%
“…Initially, DNNs were used to extract descriptors of image patches at predefined keypoints [7]. Subsequently, the mainstream approach became the extraction of keypoints and descriptors with a single network [8]- [10], which can often extract more robust keypoints and discriminative descriptors than hand-crafted methods [11]. We refer to these methods as map-based methods because they estimate a score map and a descriptor map using two heads: the score map head (SMH) and the descriptor map head (DMH).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…This step is non-differentiable, therefore we apply a differentiable subpixel keypoint refinement (DKR) that allows the gradients to flow back to the small regions around the keypoints. Inspired by recent works [14,15,11,27], we extract 5 × 5 patches from the confidence heatmap which are centered at the N max keypoint positions and compute for each patch p the spatial softargmax of the normalized patch (p − s NMS )/t, where s NMS is the value of the NMS score map and t the temperature for the softmax. The refined keypoint positions are the sum of the initial coordinates and the soft subpixel coordinates.…”
Section: Multi-modal Retinal Keypoint Detection and Description Networkmentioning
confidence: 99%