Proceedings of the 16th ACM International Conference on Multimedia 2008
DOI: 10.1145/1459359.1459392
|View full text |Cite
|
Sign up to set email alerts
|

Video event detection using motion relativity and visual relatedness

Abstract: Event detection plays an essential role in video content analysis. However, the existing features are still weak in event detection because: i) most features just capture what is involved in an event or how the event evolves separately, and thus cannot completely describe the event; ii) to capture event evolution information, only motion distribution over the whole frame is used which proves to be noisy in unconstrained videos; iii) the estimated object motion is usually distorted by camera movement. To cope w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
49
0

Year Published

2011
2011
2021
2021

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 80 publications
(49 citation statements)
references
References 31 publications
(39 reference statements)
0
49
0
Order By: Relevance
“…However, computation of trajectory descriptors requires substantial computational overhead. The first of its kind was proposed by Wang et al [149] where the authors used the wellknown Kanade-Lucas-Tomasi (KLT) tracker [79] to extract DoG-SIFT key-point trajectories, and compute a feature by modeling the motion between every trajectory pair. Sun et al [132] also applied KLT to track DoG-SIFT key-points.…”
Section: Trajectory Descriptorsmentioning
confidence: 99%
See 1 more Smart Citation
“…However, computation of trajectory descriptors requires substantial computational overhead. The first of its kind was proposed by Wang et al [149] where the authors used the wellknown Kanade-Lucas-Tomasi (KLT) tracker [79] to extract DoG-SIFT key-point trajectories, and compute a feature by modeling the motion between every trajectory pair. Sun et al [132] also applied KLT to track DoG-SIFT key-points.…”
Section: Trajectory Descriptorsmentioning
confidence: 99%
“…Different from [149], they computed three levels of trajectory context, including point-level context which is an averaged SIFT descriptor, intra-trajectory context which models trajectory transitions over time, and inter-trajectory context which encodes proximities between trajectories. The velocity histories of key-point trajectories are modeled by Messing et al [87], who observed that velocity information is useful for detecting daily living actions in high-resolution videos.…”
Section: Trajectory Descriptorsmentioning
confidence: 99%
“…For example, Wang et al [24] proposed to incorporate a number of motion primitives of each visual word into the BovW representation of videos. Nevertheless, the above approaches usually directly include the spatial-temporal information into visual content representation, and the storage and computational cost is often high.…”
Section: Related Workmentioning
confidence: 99%
“…near identical regions being assigned to di↵erent visual words, soft-quantization of visual words [9] has been proposed to map each descriptor onto multiple neighboring visual words (in the descriptor feature space). Despite its simple structure, the BovW model has shown a promising performance in the fields such as object/event recognition [24] and image/video retrieval [20].…”
Section: Introductionmentioning
confidence: 99%
“…Inspired by the method proposed by Wang et al [5] which classifies the relative motion of visual words to represent the temporal patterns in a video, we propose to utilize relative motion to model the temporal relation of visual words.…”
Section: Introductionmentioning
confidence: 99%