2005 IEEE International Conference on Multimedia and Expo
DOI: 10.1109/icme.2005.1521463
|View full text |Cite
|
Sign up to set email alerts
|

Comparing Feature Sets for Acted and Spontaneous Speech in View of Automatic Emotion Recognition

Abstract: We present a data-mining experiment on feature selection for automatic emotion recognition. Starting from more than 1000 features derived from pitch, energy and MFCC time series, the most relevant features in respect to the data are selected from this set by removing correlated features. The features selected for acted and realistic emotions are analysed and show significant differences. All features are computed automatically and we also contrast automatically with manually units of analysis. A higher degree … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
121
0
2

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 176 publications
(125 citation statements)
references
References 4 publications
2
121
0
2
Order By: Relevance
“…Literature on speech (see for example Banse and Scherer [15]) shows that the majority of studies have been conducted with emotional acted speech. Feature sets for acted and spontaneous speech have recently been compared in [16]. Generally, few acted-emotion speech databases have included speakers with several different native languages.…”
Section: Related Workmentioning
confidence: 99%
“…Literature on speech (see for example Banse and Scherer [15]) shows that the majority of studies have been conducted with emotional acted speech. Feature sets for acted and spontaneous speech have recently been compared in [16]. Generally, few acted-emotion speech databases have included speakers with several different native languages.…”
Section: Related Workmentioning
confidence: 99%
“…It aligns well with the fact that elicited corpora have mostly neutral or assets that can be considered not emotive enough [25], [36], [38], but at the same time this suggests a need for more sophisticated ways of performing the discretisation.…”
Section: Resultsmentioning
confidence: 87%
“…It has been stated that prosodic features are especially useful in the case of acted speech, but spectral features are valuable when recognising emotion from natural or elicited speech [38]. Some researchers also state the importance of lexical and contextual features [11], but their extraction usually requires the additional effort of annotating the speech corpus.…”
Section: Classifying Emotion In Speechmentioning
confidence: 99%
“…Such features include, but are not limited to, speech and its content, prosodic and paralinguistic features, eye gaze, facial expressions, body movements, or more advanced interpretations of such features such as the affective state, personality, mood or intentions of the user (e.g. [8,14,15]). …”
Section: Analysis Of Natural Interactionsmentioning
confidence: 99%