2014
DOI: 10.1109/tip.2014.2315959
|View full text |Cite
|
Sign up to set email alerts
|

Derivative-Based Scale Invariant Image Feature Detector With Error Resilience

Abstract: We present a novel scale-invariant image feature detection algorithm (D-SIFER) using a newly proposed scale-space optimal 10th-order Gaussian derivative (GDO-10) filter, which reaches the jointly optimal Heisenberg's uncertainty of its impulse response in scale and space simultaneously (i.e., we minimize the maximum of the two moments). The D-SIFER algorithm using this filter leads to an outstanding quality of image feature detection, with a factor of three quality improvement over state-of-the-art scale-invar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2015
2015
2018
2018

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(17 citation statements)
references
References 17 publications
0
17
0
Order By: Relevance
“…In addition to this operator, KAZE adds the novel concept of using a non-linear diffusion scale space instead of the more traditional Gaussian pyramid, making it the de facto current state of the art among handcrafted keypoint detectors. More recently, SIFER [14] and D-SIFER [13] proposed an advanced Cosine Modulated Gaussian filter instead of traditional derivative-based ones, with promising results.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition to this operator, KAZE adds the novel concept of using a non-linear diffusion scale space instead of the more traditional Gaussian pyramid, making it the de facto current state of the art among handcrafted keypoint detectors. More recently, SIFER [14] and D-SIFER [13] proposed an advanced Cosine Modulated Gaussian filter instead of traditional derivative-based ones, with promising results.…”
Section: Related Workmentioning
confidence: 99%
“…Applications of keypoint detection include tracking and 3D reconstruction, which often have extremely low latency and power efficiency requirements, such as in the case of autonomous driving (latency) and AR/VR pose estimation (latency and power consumption). The majority of state of the art keypoint detectors [3,12,5,14,13] are based on combinations of derivative operations, such as determinant of the Hessian [3] or difference of Gaussians [12], and their implementations are based on conventional image filtering and processing approaches. Similarly to keypoint detectors, the early layers of Convolutional Neural Networks (CNNs) are also characterized by combinations of filtering operations, hinting that keypoint detectors could be implemented as CNNs.…”
Section: Introductionmentioning
confidence: 99%
“…Some typical examples of the detectors of this type are Harris corners [4] for corner detection, SIFT [5], SURF [6], MSER [7] for blob detection, and SFOP [8] for junction detection. Besides the list of the aforementioned detectors, there are a vast number of detectors such as SIFER [11], D-SIFER [12], WADE [13], Edge Foci [2] targeting detection of different structures with various customizations. Although current detectors rely on some more or less different pre-designed structures, the structures share a common factor in that they have some levels of complexity.…”
Section: Hand-crafted Feature Detectormentioning
confidence: 99%
“…Given the spatial Gaussian scale-space concept [24,34,44,46,47,59,60,67,70,106,111,120,123], a general methodology for spatial scale selection has been developed based on local extrema over spatial scales of scale-normalized differential entities [62,64,65,72,73]. This general method- 2 The spatial Laplacian applied to the first-and second-order temporal derivatives ∇ 2 (x,y) L t and ∇ 2 (x,y) L tt as well as the spatio-temporal Laplacian ∇ 2 (x,y,t) L computed from a video sequence in the UCF-101 dataset (Kayaking_g01_c01.avi) at 3 × 3 combinations of the spatial scales (bottom row) σ s,1 = 2 pixels, (middle row) σ s,2 = 4.6 pixels and (top row) σ s,3 = 10.6 pixels and the temporal scales (left column) σ τ,1 = 40 ms, (middle column) σ τ,2 = 160 ms and (right column) σ τ,3 = 640 ms with the spatial and temporal scale parameters in units of σ s = √ s and σ τ = √ τ and using a time-causal spatio-temporal scale-space representation with a logarithmic distribution of the temporal scale levels for c = 2 (image size: 320 × 172 pixels of original 320 × 240 pixels; frame 90 of 226 frames at 25 framesframes/s) ology has in turn been successfully applied to develop robust methods for image-based matching and recognition [5,41,52,68,74,84,86,87,89,90,[112][113][114] that are able to handle large variations of the size of the objects in the image domain and with numerous applications regarding object recognition, object categorization, multi-view geometry, construction of 3-D models from visual input,…”
Section: Figmentioning
confidence: 99%