2022 26th International Conference on Pattern Recognition (ICPR) 2022
DOI: 10.1109/icpr56361.2022.9956589
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Emotion Recognition with Modality-Pairwise Unsupervised Contrastive Loss

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3

Relationship

1
5

Authors

Journals

citations
Cited by 11 publications
(10 citation statements)
references
References 61 publications
0
1
0
Order By: Relevance
“…Motivated by the successful application of CL in unsupervised learning (Oord, Li, and Vinyals 2018;He et al 2020), Supervised Contrastive Learning (SCL) (Khosla et al 2020) is devised to promote a series of supervised tasks. Recently, CL has been applied to multi-modal tasks to strengthen the interaction between features of different modalities (Zheng et al 2022;Franceschini et al 2022;Zolfaghari et al 2021). However, there has been no exploration of contrastive learning on multi-modal tasks in the multi-label scenario.…”
Section: Related Workmentioning
confidence: 99%
“…Motivated by the successful application of CL in unsupervised learning (Oord, Li, and Vinyals 2018;He et al 2020), Supervised Contrastive Learning (SCL) (Khosla et al 2020) is devised to promote a series of supervised tasks. Recently, CL has been applied to multi-modal tasks to strengthen the interaction between features of different modalities (Zheng et al 2022;Franceschini et al 2022;Zolfaghari et al 2021). However, there has been no exploration of contrastive learning on multi-modal tasks in the multi-label scenario.…”
Section: Related Workmentioning
confidence: 99%
“…In the last two decades, several researchers proposed models for automatic emotion recognition from nonverbal cues such as voice activity [1], [2], [3], body motions [4], [5], [6], touch [7], as well as their combinations [8], [9]. However, the most often considered indicators of emotional states are facial expressions [10], [11], [12], [13].…”
Section: Introductionmentioning
confidence: 99%
“…The success of FER predominantly reckons on the supervised learning paradigm in which the data annotation is expensive and laborious. Importantly, obtaining highly reliable emotion labels is tough [8] since the perception of emotional expressions depends on several factors such as gender and culture [31]. There exists a few attempts to perform unsupervised learning: Xiao et al [32] apply Restricted Boltzmann Machines (RBMs), and Yu et al [33] use Cycle Generative Adversarial Network (CycleGAN), for this purpose.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations