2002 IEEE Workshop on Multimedia Signal Processing.
DOI: 10.1109/mmsp.2002.1203271
|View full text |Cite
|
Sign up to set email alerts
|

Video genre verification using both acoustic and visual modes

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(9 citation statements)
references
References 10 publications
0
9
0
Order By: Relevance
“…Roach et al [42] extend their earlier work, classifying video using the audio features described in Roach and Mason [41] as well as visual features obtained in a manner similar to that described in Roach et al [75]. A GMM is used for classification of a linear combination of the conditional probabilities of the audio and visual features.…”
Section: Combination Approachesmentioning
confidence: 90%
See 2 more Smart Citations
“…Roach et al [42] extend their earlier work, classifying video using the audio features described in Roach and Mason [41] as well as visual features obtained in a manner similar to that described in Roach et al [75]. A GMM is used for classification of a linear combination of the conditional probabilities of the audio and visual features.…”
Section: Combination Approachesmentioning
confidence: 90%
“…Roach et al [42] detect the motion of foreground objects using a frame-differencing approach. Pixel-wise frame differencing of consecutive frames is performed using the Euclidean distance between pixels in the RGB color space.…”
Section: ) Motion-based Featuresmentioning
confidence: 99%
See 1 more Smart Citation
“…One approach is to use the output of different Hidden Markov Models as the input of a multi-layer perceptron Neural Network (Huang J. et al, 1999). Another approach makes use of a Gaussian mixture model to classify a linear combination of the conditional probabilities of audio and visual features (Roach et al, 2002). A simpler idea is to concatenate different features into a single vector that will be used to train an SVM, as for example described in (Lin and Hauptmann, 2002).…”
Section: Related Workmentioning
confidence: 99%
“…Roach et al [RMX02] detect the motion of foreground objects using a framedifferencing approach. Pixel-wise frame differencing of consecutive frames is performed using the Euclidean distance between pixels in the RGB color space.…”
Section: Motion-based Featuresmentioning
confidence: 99%