2007
DOI: 10.1049/iet-ipr:20060040
|View full text |Cite
|
Sign up to set email alerts
|

Bottom-up spatiotemporal visual attention model for video analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
34
0
1

Year Published

2009
2009
2021
2021

Publication Types

Select...
4
2

Relationship

2
4

Authors

Journals

citations
Cited by 43 publications
(35 citation statements)
references
References 37 publications
0
34
0
1
Order By: Relevance
“…Specifically, we extend the spatial center-surround operator of Itti et al in a straightforward manner by using volumetric neighborhoods in a spatiotemporal Gaussian scale-space (Rapantzikos et al 2007). In such a framework, a video sequence is treated as a volume, which is created by stacking temporally consequent video frames.…”
Section: Volumetric Saliency By Feature Competitionmentioning
confidence: 99%
See 1 more Smart Citation
“…Specifically, we extend the spatial center-surround operator of Itti et al in a straightforward manner by using volumetric neighborhoods in a spatiotemporal Gaussian scale-space (Rapantzikos et al 2007). In such a framework, a video sequence is treated as a volume, which is created by stacking temporally consequent video frames.…”
Section: Volumetric Saliency By Feature Competitionmentioning
confidence: 99%
“…This model has proven its efficiency in enhancing performance of a video classification system (Rapantzikos and Avrithis 2005). The two other saliency-based methods are the state-of-the art static saliency-based approach of Itti et al (1998) and an extension using a motion map (Rapantzikos and Tsapatsoulis 2005).…”
Section: Evaluation Of Classification Performancementioning
confidence: 99%
“…The different orientations are then fused to produce a single orientation volume. More details can be found in [3]. Volumes for each feature, are decomposed into multiple scales.…”
Section: Visual Analysismentioning
confidence: 99%
“…Perceptual attention is triggered by changes in the involved events like scene transitions, progressions or newly introduced themes. Computational models of attention have been previously developed using multimodal analysis, i.e., the concurrent analysis of multiple information modalities [1,2,3,4,5]. Automatic video content access, analysis and abstraction have thus emerged as potential applications.…”
Section: Introductionmentioning
confidence: 99%
“…Visual input is processed by computer vision [15] (see Section 4.1) or synthetic vision techniques [2], as appropriate, and stored in a short-term sensory storage. This acts as a temporary buffer and contains a large amount of raw data for short periods of time.…”
Section: General Frameworkmentioning
confidence: 99%