Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-2464
|View full text |Cite
|
Sign up to set email alerts
|

Predicting Categorical Emotions by Jointly Learning Primary and Secondary Emotions through Multitask Learning

Abstract: Detection of human emotions is an essential part of affect-aware human-computer interaction (HCI). In daily conversations, the preferred way of describing affects is by using categorical emotion labels (e.g., sad, anger, surprise). In categorical emotion classification, multiple descriptors (with different degrees of relevance) can be assigned to a sample. Perceptual evaluations have relied on primary and secondary emotions to capture the ambiguous nature of spontaneous recordings. Primary emotion is the most … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
28
0

Year Published

2020
2020
2025
2025

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 36 publications
(28 citation statements)
references
References 21 publications
(26 reference statements)
0
28
0
Order By: Relevance
“…They base their experiments on three datasets for evaluations and demonstrate that the proposed MTRL model gains a concordance correlation coefficient (CCC) as high as 4.7% for within-corpus and 14.0% for cross-corpora experiments compared to STL. In [112], the authors introduce an MTRL framework for jointly learning primary and secondary emotions. They perform evaluations on the MSP-Podcast database and show that the proposed MTRL model can leverage the extra information about the secondary emotions and leads to relative improvements of 7.9% in F1-score for an 8-class emotion classification task.…”
Section: Multi-task Representation Learningmentioning
confidence: 99%
“…They base their experiments on three datasets for evaluations and demonstrate that the proposed MTRL model gains a concordance correlation coefficient (CCC) as high as 4.7% for within-corpus and 14.0% for cross-corpora experiments compared to STL. In [112], the authors introduce an MTRL framework for jointly learning primary and secondary emotions. They perform evaluations on the MSP-Podcast database and show that the proposed MTRL model can leverage the extra information about the secondary emotions and leads to relative improvements of 7.9% in F1-score for an 8-class emotion classification task.…”
Section: Multi-task Representation Learningmentioning
confidence: 99%
“…Since the exact position of an adjective in this space may be imprecise [21], we reduced the possibilities by defining four categories of quadrants following [22]: Q1 (positive arousal -positive valence), Q2 (positive arousal -negative valence), Q3 (negative arousal -negative valence), and Q4 (negative arousal -positive valence). By employing multi-task learning (MTL), our classifier was trained simultaneously for four-class classification (in the case of quadrants) and for binary classification (positive and negative arousal and valence) in order to improve generalization [23,24]. Our results showed feasibility of our hypothesis: intra-linguistic settings (e.g., pretraining on speech in English and fine-tuning with music in English) yield better results than cross-linguistic settings (e.g., pretraining on speech in Mandarin and fine-tuning with music in English).…”
Section: Related Workmentioning
confidence: 99%
“…Moreover, this subset is strongly dependent of the task. For instance, a political debate on TV [16] or a podcast interview [17] cannot be expected to show the same human emotions than a human-machine scenario [10], [18], [19]. Indeed, fear is not expected in none of the mentioned corpus.…”
Section: Introductionmentioning
confidence: 99%