2010
DOI: 10.1007/s10772-010-9068-y
|View full text |Cite
|
Sign up to set email alerts
|

Emotion recognition and adaptation in spoken dialogue systems

Abstract: The involvement of emotional states in intelligent spoken human-computer interfaces has evolved to a recent field of research. In this article we describe the enhancements and optimizations of a speech-based emotion recognizer jointly operating with automatic speech recognition. We argue that the knowledge about the textual content of an utterance can improve the recognition of the emotional content. Having outlined the experimental setup we present results and demonstrate the capability of a postprocessing al… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
39
0
2

Year Published

2011
2011
2022
2022

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 56 publications
(41 citation statements)
references
References 20 publications
(13 reference statements)
0
39
0
2
Order By: Relevance
“…An emotion recognizer that may operate jointly with an automatic speech recognizer is examined by Pitterman et al [46]. The feature vector comprises of MFCCs (along with their first-and second-order differences), intensity, and three formants, along with pitch and pitch statistics, namely minimum, mean, maximum, deviation and range.…”
Section: Emotion Recognition On Emodbmentioning
confidence: 99%
See 2 more Smart Citations
“…An emotion recognizer that may operate jointly with an automatic speech recognizer is examined by Pitterman et al [46]. The feature vector comprises of MFCCs (along with their first-and second-order differences), intensity, and three formants, along with pitch and pitch statistics, namely minimum, mean, maximum, deviation and range.…”
Section: Emotion Recognition On Emodbmentioning
confidence: 99%
“…No feature selection technique is applied, while the HMMs are employed as classifiers to a speaker-dependent protocol, contrary to our approach that applies feature selection and a speaker-independent protocol. Also, speech recognition is not a prerequisite in this work, whereas the stated set of features in [46] is a subset of the feature vector computed by the authors. However, in both cases the authors compute the first-and second-differences of the features in order to capture their temporal evolution.…”
Section: Emotion Recognition On Emodbmentioning
confidence: 99%
See 1 more Smart Citation
“…The dialog model proposed by [121] combined three different submodels: an emotional model describing the transitions between user emotional states during the interaction regardless of the data content, a plain dialog model describing the transitions between existing dialog states regardless of the emotions, and a combined model including the dependencies between combined dialog and emotional states. Then, the next dialog state was derived from a combination of the plain dialog model and the combined model.…”
Section: Modeling the User Emotional Statementioning
confidence: 99%
“…In our proposal, we employ statistical techniques for inferring user acts, which makes it easier to port it to different application domains. Also the proposed architecture is modular and thus makes it possible to employ different emotion and intention recognizers, as the intention recognizer is not linked to the dialog manager as in [121].…”
Section: Modeling the User Emotional Statementioning
confidence: 99%