2009 IEEE Conference on Computer Vision and Pattern Recognition 2009
DOI: 10.1109/cvpr.2009.5206525
|View full text |Cite
|
Sign up to set email alerts
|

Dense saliency-based spatiotemporal feature points for action recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
52
0
2

Year Published

2010
2010
2021
2021

Publication Types

Select...
5
4

Relationship

2
7

Authors

Journals

citations
Cited by 118 publications
(58 citation statements)
references
References 15 publications
(16 reference statements)
2
52
0
2
Order By: Relevance
“…Willems et al propose a space-time detector based on the determinant of the 3D Hessian matrix, which is computationally efficient (use of integral videos) and is still on a par with current methods. Quite recently, Rapantzikos et al [9] proposed and evaluated the potential of spatiotemporal saliency to represent and detect actions in video sequences. Their method ranks among the best on standard human action datasets.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Willems et al propose a space-time detector based on the determinant of the 3D Hessian matrix, which is computationally efficient (use of integral videos) and is still on a par with current methods. Quite recently, Rapantzikos et al [9] proposed and evaluated the potential of spatiotemporal saliency to represent and detect actions in video sequences. Their method ranks among the best on standard human action datasets.…”
Section: Related Workmentioning
confidence: 99%
“…The competition is implemented through constrained minimization with the constraints being inspired by the Gestalt laws. From the mathematical point of view, the computational model we use is the same with the one we proposed in [9]. In this manuscript, we provide a detailed walkthrough of the model, highlight the main principles, and give a better insight of the method.…”
Section: Introductionmentioning
confidence: 99%
“…To begin, we will start developing our theory for spatiotemporal scale selection with respect to the problem of detecting sparse spatio-temporal interest points [6,9,11,14,18,20,21,30,49,88,94,97,99,100,107,122,124,126,127], which may be regarded as a conceptually simplest problem domain because of the sparsity of spatio-temporal interest points and the close connection between this problem domain and the detection of spatial interest points for which there exists a theoretically well-founded and empirically tested framework regarding scale selection over the spatial domain [1,4,5,15,17,25,42,65,72,74,84,89,90,112]. Specifically, using a non-causal Gaussian spatio-temporal scale-space model, we will perform a theoretical analysis of the spatio-temporal scale selection properties of eight different types of spatiotemporal interest point detectors and show that seven of them: (i) the spatial Laplacian of the first-order temporal derivative, (ii) the spatial Laplacian of the second-order temporal derivative, (iii) the determinant of the spatial Hessian of the first-order temporal derivative, (iv) the determinant of the spatial Hessian of the second-order temporal derivative, (v) the determinant of the spatio-temporal Hessian matrix, (vi) the first-order temporal derivative of the determinant of the spatial Hessian matrix and (vii) the second-order temporal derivative of the determinant of the spatial Hessian matrix, do all lead to fully scale-covariant spatio-temporal scale estimates and scale-invariant feature responses under independent scaling transformations of the spatial and the temporal domains.…”
Section: Fig 4 the First-and Second-order Temporal Derivatives Of Thmentioning
confidence: 99%
“…The detector, called Cuboid, finds local maxima of a response function that contains a 2D Gaussian smoothing kernel and 1D temporal Gabor filters. Rapantzikos et al [112] used saliency to locate spatio-temporal points, where the saliency is computed by a global minimization process which leverages spatial proximity, scale, and feature similarity. To compute the feature similarity, they also utilized color, in addition to intensity and motion that are commonly adopted in other detectors.…”
Section: Spatio-temporal Local Featuresmentioning
confidence: 99%