Proceedings of the 15th ACM International Conference on Multimedia 2007
DOI: 10.1145/1291233.1291280
|View full text |Cite
|
Sign up to set email alerts
|

Practical elimination of near-duplicates from web video search

Abstract: Current web video search results rely exclusively on text keywords or user-supplied tags. A search on typical popular video often returns many duplicate and near-duplicate videos in the top results. This paper outlines ways to cluster and filter out the nearduplicate video using a hierarchical approach. Initial triage is performed using fast signatures derived from color histograms. Only when a video cannot be clearly classified as novel or nearduplicate using global signatures, we apply a more expensive local… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

5
336
0
4

Year Published

2008
2008
2019
2019

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 306 publications
(345 citation statements)
references
References 37 publications
5
336
0
4
Order By: Relevance
“…The dictionary is generated by clustering 1,185,698 keypoints extracted from 3,000 keyframes randomly sampled from the dataset of [1]. We employ DoG [6] for keypoint detection and PSIFT [9] for feature description.…”
Section: Visual Keywordmentioning
confidence: 99%
See 2 more Smart Citations
“…The dictionary is generated by clustering 1,185,698 keypoints extracted from 3,000 keyframes randomly sampled from the dataset of [1]. We employ DoG [6] for keypoint detection and PSIFT [9] for feature description.…”
Section: Visual Keywordmentioning
confidence: 99%
“…A rough statistic, as indicated in [1], shows that more than 65, 000 videos have been uploaded on video sharing web site YouTube daily. It is believed that this number is still increasing with fast speed.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In [15], near-duplicate frames are identified in the TRECVID 2004 video corpus, but 16 seconds are required for searching 150 frames in 10 minutes of video (600 frames). An application of CBVCD to the elimination of video duplicates in Web search is proposed in [14]; global descriptors help separating the least similar videos, then local descriptors allow to refine duplicate detection. However, several minutes are required for returning the top 10 answers (among the 600 preliminary results) to a keyword-based query, which is too long for performing the query-dependent online processing we need.…”
Section: Content-based Video Copy Detection For Video Miningmentioning
confidence: 99%
“…Therefore, an important problem now faced by these video sharing sites is how to automatically perform accurate and fast similarity search for an incoming video clip against its huge database, to avoid copyright violation. Meanwhile, since the retrieval efficiency will be hampered if a large number of search results are essentially almost-identical, database purge also contributes to high-quality ranking for video search results [34].…”
Section: Introductionmentioning
confidence: 99%