2008
DOI: 10.1007/978-3-540-88688-4_48
|View full text |Cite
|
Sign up to set email alerts
|

An Efficient Dense and Scale-Invariant Spatio-Temporal Interest Point Detector

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
540
0
9

Year Published

2010
2010
2022
2022

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 696 publications
(551 citation statements)
references
References 14 publications
2
540
0
9
Order By: Relevance
“…Some of the most popular descriptors have been based on the gradient of the appearance information (Schuldt et al 2004;Laptev and Perez 2007), spatio-temporal extensions to SIFT and SURF descriptors (Scovanner et al 2007;Willems et al 2008) and 2D motion information (Dalal et al 2006;Messing et al 2009). However, there has been little previous work on feature descriptors including depth information (which is generally encoded directly at the holistic level, with the aid of user masks).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Some of the most popular descriptors have been based on the gradient of the appearance information (Schuldt et al 2004;Laptev and Perez 2007), spatio-temporal extensions to SIFT and SURF descriptors (Scovanner et al 2007;Willems et al 2008) and 2D motion information (Dalal et al 2006;Messing et al 2009). However, there has been little previous work on feature descriptors including depth information (which is generally encoded directly at the holistic level, with the aid of user masks).…”
Section: Related Workmentioning
confidence: 99%
“…This can be done at various stages of the pipeline, but we first look at interest point extraction, in order to detect more salient features, and discount irrelevant detections. The extended algorithms discussed in this section are based on the Harris Corners work by Laptev and Lindeberg (2003), the Hessian points algorithm by Willems et al (2008) and the Separable Filters technique by Dollar et al (2005). For a comparison of the original 2D interest point detection schemes (without the proposed depth-aware extensions), see the survey paper by Tuytelaars and Mikolajczyk (2008).…”
Section: Interest Point Detectionmentioning
confidence: 99%
“…A cal-ibrated camera moves freely in two scenarios: multiple objects with and without repetition. To evaluate the performance of our system, a recognition is deemed valid if J > 0.5, as proposed in [19]. In the first scenario, five different objects are present in the scene, while the second scenario envisages four instances of the same object and two other objects.…”
Section: Multiple Object Recognitionmentioning
confidence: 99%
“…The BoF approach can also be extended to action recognition tasks. The most direct way to do so is to build a histogram for each video, where the features are extracted from groups of frames rather than from a single image (see, e.g., [21][22][23]). In the case of surgical gesture recognition, we extract Space-Time Interest Points (STIP) [21] from each video surgeme.…”
Section: Classification Using Bag Of Spatio-temporal Featuresmentioning
confidence: 99%