IFIP the International Federation for Information Processing
DOI: 10.1007/978-0-387-74161-1_41
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal emotion recognition from expressive faces, body gestures and speech

Abstract: Abstract. In this paper we present a multimodal approach for the recognition of eight emotions that integrates information from facial expressions, body movement and gestures and speech. We trained and tested a model with a Bayesian classifier, using a multimodal corpus with eight emotions and ten subjects. First individual classifiers were trained for each modality. Then data were fused at the feature level and the decision level. Fusing multimodal data increased very much the recognition rates in comparison … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
49
0
5

Publication Types

Select...
4
4
2

Relationship

2
8

Authors

Journals

citations
Cited by 112 publications
(56 citation statements)
references
References 23 publications
2
49
0
5
Order By: Relevance
“…Their study shows that sadness and anger are detected more easily from speech, while the recognition of joy and fear is less reliable. Caridakis et al [34] obtained 93.30% and 76.67% accuracy to identify anger and sadness, respectively, from speech, using 377 features based on intensity, pitch, Mel-Scale Frequency Cepstral Coefficients (MFCC), Bark spectral bands, voiced segment characteristics, and pause length.…”
Section: Methodsmentioning
confidence: 99%
“…Their study shows that sadness and anger are detected more easily from speech, while the recognition of joy and fear is less reliable. Caridakis et al [34] obtained 93.30% and 76.67% accuracy to identify anger and sadness, respectively, from speech, using 377 features based on intensity, pitch, Mel-Scale Frequency Cepstral Coefficients (MFCC), Bark spectral bands, voiced segment characteristics, and pause length.…”
Section: Methodsmentioning
confidence: 99%
“…In these datasets, interactions include two interlocutors, who are recorded carrying out both structured and unstructured conversations. ere are numerous examples of such datasets, with a wide range of applications, such as speech recognition [15], behavior analysis [50], segmentation, emotion recognition [12] and depression detection [16]. Arguably, one of the most popular datasets of one-to-one interactions is SEMAINE [30].…”
Section: Related Workmentioning
confidence: 99%
“…The authors relied on the body movement to recognize the emotion instead of focusing on the overall body posture as a whole. Additionally, Caridakis et al proposed a method which utilized a Bayesian classifier to recognize emotion based on body gestures and speech [11]. In this literature, the authors collected multimodal data and then trained the data with a Bayesian classifier to recognize the emotion.…”
Section: Related Workmentioning
confidence: 99%