2007
DOI: 10.1109/tmm.2007.893337
|View full text |Cite
|
Sign up to set email alerts
|

Enhanced Eigen-Audioframes for Audiovisual Scene Change Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0
1

Year Published

2008
2008
2015
2015

Publication Types

Select...
5
2
2

Relationship

1
8

Authors

Journals

citations
Cited by 17 publications
(7 citation statements)
references
References 25 publications
0
6
0
1
Order By: Relevance
“…This may be particularly useful in some real life applications such as movie/TV data indexing/retrieval, annotating media data through audio labels, and media content understanding. Figure 4 further compares the performance of HMM with the sliding window method [7] in automatically labeling audio events, from which we can see that the HMM model achieves an averagely higher accuracy than the sliding window method. …”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…This may be particularly useful in some real life applications such as movie/TV data indexing/retrieval, annotating media data through audio labels, and media content understanding. Figure 4 further compares the performance of HMM with the sliding window method [7] in automatically labeling audio events, from which we can see that the HMM model achieves an averagely higher accuracy than the sliding window method. …”
Section: Resultsmentioning
confidence: 99%
“…For example, Chu et al [6] perform an empirical feature analysis for audio environment characterization and propose to use the matching pursuit (MP) algorithm to obtain effective timefrequency features. Kyperountas et al [7] create an enhanced set of eigen-audioframes that is related to an audio signal subspace to discover audio background changes. Unfortunately, how to bridge the semantic gap that separates low-level auditory features and high-level auditory contents is still difficult even by combining more heterogeneous features.…”
Section: Introductionmentioning
confidence: 99%
“…References [8][9][10] combined audio and video information as "audiovisual feature" to jointly solve the scene change detection problem, improved the detection accuracy in some degrees but also brought additional computations.…”
Section: Related Workmentioning
confidence: 99%
“…On a news programme, we assume that the whole presentation of a piece of news is a unique scene, once all related shots talk about the news' subject. The definition of the beginning and ending timestamps of each scene is done manually, because this process is out of scope in this work; however, automatic algorithms have been reported on literature [21,23]. Following this process, we generate a set of SubRip files (SRT), using the open-source closed-caption tool CCExtractor 7 , which converts binary closed-caption captured with Microsoft GraphEdit 8 from analog sources into subtitle text.…”
Section: Indexing and Enrichment: Backgroundmentioning
confidence: 99%