Procedings of the British Machine Vision Conference 2010 2010
DOI: 10.5244/c.24.95
|View full text |Cite
|
Sign up to set email alerts
|

Improving bag-of-features action recognition with non-local cues

Abstract: Local space-time features have recently shown promising results within Bag-of-Features (BoF) approach to action recognition in video. Pure local features and descriptors, however, provide only limited discriminative power implying ambiguity among features and sub-optimal classification performance. In this work, we propose to disambiguate local space-time features and to improve action recognition by integrating additional nonlocal cues with BoF representation. For this purpose, we decompose video into region … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
75
0

Year Published

2013
2013
2021
2021

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 92 publications
(76 citation statements)
references
References 23 publications
(35 reference statements)
1
75
0
Order By: Relevance
“…Table 2 shows that our approach outperforms all other methods reported in literature so far. In particular, results are better than those of Ullah et al [7], who used a person detector, and Vig et al [14], who built upon saliency from an eye tracking system. A detailed analysis of the contribution of each set of features is presented in Table 3.…”
Section: Methodsmentioning
confidence: 45%
See 4 more Smart Citations
“…Table 2 shows that our approach outperforms all other methods reported in literature so far. In particular, results are better than those of Ullah et al [7], who used a person detector, and Vig et al [14], who built upon saliency from an eye tracking system. A detailed analysis of the contribution of each set of features is presented in Table 3.…”
Section: Methodsmentioning
confidence: 45%
“…However, such failure cases are quite rare, and they are outnumbered by those cases where current person detectors fail. Hence, compared to [7], we achieve much better performance. We also compare to [14], where an eye tracking system was used to emphasize the part of the image that humans consider most important.…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations