2013
DOI: 10.1016/j.neucom.2012.03.034
|View full text |Cite
|
Sign up to set email alerts
|

Automatic highlights extraction for drama video using music emotion and human face features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(5 citation statements)
references
References 16 publications
0
5
0
Order By: Relevance
“…A system using music emotion and the human face as features for drama video has been proposed. 13 The method concerned uses two high-level features (music emotion and the human face) and two low-level features (shot duration and motion magnitude) to extract highlights. The authors claim that the method is effective for video highlight extraction.…”
Section: Related Workmentioning
confidence: 99%
“…A system using music emotion and the human face as features for drama video has been proposed. 13 The method concerned uses two high-level features (music emotion and the human face) and two low-level features (shot duration and motion magnitude) to extract highlights. The authors claim that the method is effective for video highlight extraction.…”
Section: Related Workmentioning
confidence: 99%
“…First, following the definition in previous work (M. Xu, Jin, Luo, & Duan, 2008), we define highlights as the most memorable shots in a video with high emotion intensity. Note that highlight detection is different from video summarization, which focuses on condensed storyline representation of a video, rather than extracting affective contents (K.-S. Lin, Lee, Yang, Lee, & Chen, 2013).…”
Section: Highlight Detection By Video Processingmentioning
confidence: 99%
“…For highlight detection, some researchers propose to represent emotions in a video by a curve on the arousal-valence plane with low-level features such as motion, vocal effects, shot length, and audio pitch (Hanjalic & Xu, 2005), color (Ngo, Ma, & Zhang, 2005), mid-level features such as laughing and subtitles (M. Xu, Luo, Jin, & Park, 2009). Nevertheless, due to the semantic gap between low-level features and high-level semantics, accuracy of highlight detection based on video processing is limited (K.-S. Lin et al, 2013).…”
Section: Highlight Detection By Video Processingmentioning
confidence: 99%
“…Moreover, other studies rely on approaches that are based on the exploitation of audio by detecting any sudden variation of sound recording [8]- [10] caused by audience reaction, such as applauds or shouts out, this variation reflect the highlight moment during the event (Fig. 2), but unfortunately this approach can have several inconveniences such as special effects added during the event, or even the presence of noise during the whole event.…”
Section: Introductionmentioning
confidence: 99%