2004
DOI: 10.1016/j.cag.2004.03.002
|View full text |Cite
|
Sign up to set email alerts
|

Revealing the connoted visual code: a new approach to video classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2010
2010
2012
2012

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 6 publications
0
5
0
Order By: Relevance
“…Each cluster representative (typically the centroid) is considered as a visual word of the visual dictionary. The K-means clustering algorithm [1,4] is the most common method to create such visual dictionaries even though other unsupervised methods such as K-median clustering [5], mean-shift clustering [6], hierarchical K-means [7], agglomerative clustering [8], radius basedclustering [6,9], or regular lattice-based strategies [10] have also been used. One of the common features of these unsupervised methods is that they only optimize an objective function fitting to the data but ignoring their class information.…”
Section: Introductionmentioning
confidence: 99%
“…Each cluster representative (typically the centroid) is considered as a visual word of the visual dictionary. The K-means clustering algorithm [1,4] is the most common method to create such visual dictionaries even though other unsupervised methods such as K-median clustering [5], mean-shift clustering [6], hierarchical K-means [7], agglomerative clustering [8], radius basedclustering [6,9], or regular lattice-based strategies [10] have also been used. One of the common features of these unsupervised methods is that they only optimize an objective function fitting to the data but ignoring their class information.…”
Section: Introductionmentioning
confidence: 99%
“…e.g., [17]. Also related, is work on video classification that makes use of connoted visual codes [18]. The output of the approach is a class label that can be used as tag for annotating the video.…”
Section: Tags Derived From Audiovisual Contentmentioning
confidence: 99%
“…Generally, cluster centers are considered as visual words. They are usually extracted by K-means based-algorithms [1,3] even though other approaches have been applied, such as k-median clustering [4], mean-shift clustering [5], hierarchi-cal K-means [6], agglomerative clustering [7], randomized trees [8], radius based-clustering [5,9], or regular lattice-base strategies [10]. Thanks to the vector quantization, a given image can then be mapped into this new space of visual words leading to a bag of visual words, where each word can be weighted either according to its frequency or using more sophisticated techniques.…”
Section: Introductionmentioning
confidence: 99%