Interspeech 2012 2012
DOI: 10.21437/interspeech.2012-118
|View full text |Cite
|
Sign up to set email alerts
|

Emotion recognition using acoustic and lexical features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 35 publications
(9 citation statements)
references
References 0 publications
0
9
0
Order By: Relevance
“…Early fusion is the fusion approach used in the pre-extraction phase of the data. Rozgic et al [ 18 ] used early fusion to connect multimodal representations as input to an inference model, which provides a novel idea for modal fusion. Zadeh et al [ 19 ] designed a memory fusion network (MFN) using multiview sequential learning, which explicitly illustrates two interactions in the neural architecture.…”
Section: Related Workmentioning
confidence: 99%
“…Early fusion is the fusion approach used in the pre-extraction phase of the data. Rozgic et al [ 18 ] used early fusion to connect multimodal representations as input to an inference model, which provides a novel idea for modal fusion. Zadeh et al [ 19 ] designed a memory fusion network (MFN) using multiview sequential learning, which explicitly illustrates two interactions in the neural architecture.…”
Section: Related Workmentioning
confidence: 99%
“…Due to their ubiquity, most works on multimodal emotion recognition have focused on combining audio and video [56,66], but how to combine them remains an open question. In early fusion, inputs or raw feature representations are merged before they are fed into a joint network [17,57]. In modellevel fusion, each modality is processed by a dedicated network before both intermediate feature representations are merged and then passed through a joint network [16,54].…”
Section: Multimodal Emotion Recognitionmentioning
confidence: 99%
“…For modality aggregation, Viktor et al [36] use early fusion to concatenate multi-modal features as the input for the inference models. But it ignores the mismatch between different modalities.…”
Section: Related Work 21 Multi-modal Emotion Recognitionmentioning
confidence: 99%
“…Such rich information from multimodalities could be used to understand the emotional state [29]. Previous research works have shown that different modalities are complementary for emotion recognition [23,36]. Different modalities all carry emotion relevant information and how to effectively combine multiple modalities has been an active research focus.…”
Section: Introductionmentioning
confidence: 99%