2015
DOI: 10.48550/arxiv.1503.01800
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

EmoNets: Multimodal deep learning approaches for emotion recognition in video

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 0 publications
0
4
0
Order By: Relevance
“…Multimodal networks have been proposed in both unsupervised ( [30]) and supervised ( [18], [22]) settings. Both [18] and [22] first train deep models on individual data modalities then use activations from these models to train a multimodal classifier.…”
Section: B Neural Networkmentioning
confidence: 99%
“…Multimodal networks have been proposed in both unsupervised ( [30]) and supervised ( [18], [22]) settings. Both [18] and [22] first train deep models on individual data modalities then use activations from these models to train a multimodal classifier.…”
Section: B Neural Networkmentioning
confidence: 99%
“…To the authors' knowledge, the only works that previously applied CNNs to expression data were that of Kahou et al [13,12] and Jung et al [11]. In [13,12], the authors developed a system for doing audio/visual emotion recognition for the Emotion Recognition in the Wild Challenge (EmotiW) [6,5] while in [11], the authors trained a network that incorporated both appearance and geometric features when doing recognition. However, one key point is that these works dealt with emotion recognition of video / image sequence data and therefore, actively incorporated temporal data when computing their predictions.…”
Section: Related Workmentioning
confidence: 99%
“…Therefore, the research focused strongly on face recognition, an active research area in recent years. In this category are the works proposed by Kahou et al 2015, Kollias et al 2015and Wei et al 2017. Unfortunately, these solutions lose generality as they are strongly focused on the primary detection of the face without considering other aspects that make up the image.…”
Section: Color and Emotion: From Computing To Deep Learningmentioning
confidence: 99%