2009
DOI: 10.1007/978-3-642-00525-1_26
|View full text |Cite
|
Sign up to set email alerts
|

Recognition of Emotions in German Speech Using Gaussian Mixture Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
6
0

Year Published

2010
2010
2022
2022

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(8 citation statements)
references
References 5 publications
1
6
0
Order By: Relevance
“…It is not entirely consistent with the results obtained from other authors using the EMO-DB database for GMM emotion recognition [37][38][39] as well as those published in more complex comparison studies [40,41]. Usually, the best recognized emotions are anger and sadness followed by neutral state, the emotion joy generates the most confusion being recognized as anger [39]. Similar results were also achieved in classifications accomplished in [33], where the same emotional speech database was used.…”
Section: Discussion Of Resultssupporting
confidence: 82%
“…It is not entirely consistent with the results obtained from other authors using the EMO-DB database for GMM emotion recognition [37][38][39] as well as those published in more complex comparison studies [40,41]. Usually, the best recognized emotions are anger and sadness followed by neutral state, the emotion joy generates the most confusion being recognized as anger [39]. Similar results were also achieved in classifications accomplished in [33], where the same emotional speech database was used.…”
Section: Discussion Of Resultssupporting
confidence: 82%
“…The most commonly used spectral features for emotion recognition are Mel-Frequency Cepstral Coefficients (MFCC) [35, 16, 15, 22, 29, 19, 11, 37, 8, 30, 21, 27, 14, 12, 32, 33, 36, 38, 39]. As in automatic speech recognition, MFCC are extracted using a 25 ms Hamming window at intervals of 10 ms and cover frequency range from 300 Hz to the Nyquist frequency.…”
Section: Prior Workmentioning
confidence: 99%
“…A real-time systems for discriminating between angry and neutral speech was implemented in [14] using GMMs for MFCC features in combination with a prosody-based classifier. Vondra and Vich [38] applied GMMs to emotion recognition using a combined feature set obtained by concatenating MFCC and prosodic features. Hu et al [12] employed the GMM supervector approach in order to extract fixed length feature vectors from utterances with variable durations.…”
Section: Prior Workmentioning
confidence: 99%
“…A variety of pattern recognition methods have been explored for automatic emotion recognition such as gaussian mixture models (Luengo et al, 2005; Vlasenko et al, 2007; Vondra and Vich, 2009), hidden Markov models (Nwe et al, 2003; Shafran et al, 2003; Meng et al, 2007), neural network (Nicholson et al, 2000) and support vector machines (Kwon et al, 2003; Tabatabaei et al, 2007; Bitouk et al, 2010; Lee et al, 2011), regression (Grimm et al, 2007b). All of these seemingly diverse methods are designed to predict the emotion of a single test utterance in isolation.…”
Section: Introductionmentioning
confidence: 99%