2008
DOI: 10.1016/j.neucom.2007.07.041
|View full text |Cite
|
Sign up to set email alerts
|

A robust multimodal approach for emotion recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
31
0
1

Year Published

2008
2008
2015
2015

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 46 publications
(32 citation statements)
references
References 14 publications
0
31
0
1
Order By: Relevance
“…Finally, combining (12), (7), and (5) into (4) yields (13) for emotional class w recognition using TCM-EWCCM:…”
Section: A Model Derivationmentioning
confidence: 99%
“…Finally, combining (12), (7), and (5) into (4) yields (13) for emotional class w recognition using TCM-EWCCM:…”
Section: A Model Derivationmentioning
confidence: 99%
“…Busso et al [2], presents a system for recognizing emotions through facial expressions displayed in live video streams and video sequences. However, works such as those presented in [3] suggest that developing a good methodology of emotional states characterization based on facial expressions, leads to more robust recognition systems. Facial Action Coding System (FACS) proposed by Ekman et al [1], is a comprehensive and anatomically based system that is used to measure all visually discernible facial movements in terms of atomic facial actions called Action Units (AUs).…”
Section: Introductionmentioning
confidence: 99%
“…As AUs are independent of interpretation, they can be used for any high-level decision-making process, including the recognition of basic emotions according to Emotional FACS (EM-FACS), the recognition of various affective states according to the FACS Affect Interpretation Database (FACSAID) introduced by Ekman et al [4], [5], [6], [7]. From the detected features, it is possible to estimate the emotion present in a particular subject, based on an analysis of estimated facial expression shape in comparision to a set of facial expressions of each emotion [3], [8], [9].…”
Section: Introductionmentioning
confidence: 99%
“…In [10] and [11], prosody features from pitch and RMS energy are extracted as audio features, and the local preserving projection (LPP) of facial texture images are adopted as visual features. To reasonably integrate the information from speech and facial expressions, [12] proposes a tripled hidden Markov model (triple-HMM) based emotion recognition system modeling the correlation of three component HMMs that are based individually on upper face, lower face, prosodic dynamic behaviors, and allows unlimited state asynchrony between these streams. Moreover, in [10 -12] it has been showed that the audio visual multi-modal emotion recognition performs much better than the mono-modal emotion recognition, i.e.…”
Section: Introductionmentioning
confidence: 99%