2009 IEEE International Conference on Signal and Image Processing Applications 2009
DOI: 10.1109/icsipa.2009.5478723
|View full text |Cite
|
Sign up to set email alerts
|

Automatic semantic video annotation in wide domain videos based on similarity and commonsense knowledgebases

Abstract: Abstract-In this paper, we introduce a novel framework for automatic Semantic Video Annotation. As this framework detects possible events occurring in video clips, it forms the annotating base of video search engine. To achieve this purpose, the system has to able to operate on uncontrolled wide-domain videos. Thus, all layers have to be based on generic features.This framework aims to bridge the "semantic gap", which is the difference between the low-level visual features and the human's perception, by findin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2009
2009
2018
2018

Publication Types

Select...
3
2
1

Relationship

3
3

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 25 publications
0
9
0
Order By: Relevance
“…There are some pioneering works in [1] and [2] concentrating on generating sentences for videos. [1] introduces a novel twostep framework for textually annotating unconstrained videos: visual similarity video matching at first and then an annotation analysis that employs commonsense knowledge bases.…”
Section: Introductionmentioning
confidence: 99%
“…There are some pioneering works in [1] and [2] concentrating on generating sentences for videos. [1] introduces a novel twostep framework for textually annotating unconstrained videos: visual similarity video matching at first and then an annotation analysis that employs commonsense knowledge bases.…”
Section: Introductionmentioning
confidence: 99%
“…A preliminary version of this work was published in conference-form in [2]. This consolidated version is extended by enhancing the first layer, introducing more technical details, and by performing more experiments on common public databases, with deeper analysis and evaluation, using standard TRECVID measures.…”
Section: Related Workmentioning
confidence: 99%
“…In the comparison phase, i.e. the distance measured between the query and each dataset's files, the video signatures are compared, and 2 An example for the framework: an airplane is taking off. The first similar video retrieved is a false positive as it is a car.…”
Section: Layer 1: Visual Similaritymentioning
confidence: 99%
“…Then ConceptNet is used to calculate the distance between the concepts. In addition to that, in our previous work [9], a full automated framework for semantic video annotation in wide-domain has been presented based on using WordNet and ConceptNet separately.…”
Section: Previous Workmentioning
confidence: 99%