Proceedings of the 2022 International Conference on Multimodal Interaction 2022
DOI: 10.1145/3536221.3556584
|View full text |Cite
|
Sign up to set email alerts
|

Supervised Contrastive Learning for Affect Modelling

Abstract: Figure 1: A high-level visualisation of the concept introduced. Supervised contrastive learning operates by infusing affect information within the representation, by pairing positive embeddings and dissociating negative embeddings. We assume affect is embedded in the multimodal latent space and defines what distinguishes (contrasts) data. Positive (green) and negative (red) multimodal data is labelled with respect to an anchor affect (black). Similar and dissimilar affect patterns define, respectively, positiv… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(12 citation statements)
references
References 45 publications
0
10
0
Order By: Relevance
“…CL is a self-supervised learning method to project the data into a space where different views of the same input sample have highly similar representations. While initially developed for image classification [54,58] and computer vision applications [59], CL has also been applied to emotion recognition based on physiological signals [60][61][62]. Mohsenvand et al [60] used a similar method to SimCLR, which they called SeqCLR, to learn similarities between differently augmented transforms of the same EEG data sample, disregarding the emotional state of the data sample.…”
Section: Contrastive Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…CL is a self-supervised learning method to project the data into a space where different views of the same input sample have highly similar representations. While initially developed for image classification [54,58] and computer vision applications [59], CL has also been applied to emotion recognition based on physiological signals [60][61][62]. Mohsenvand et al [60] used a similar method to SimCLR, which they called SeqCLR, to learn similarities between differently augmented transforms of the same EEG data sample, disregarding the emotional state of the data sample.…”
Section: Contrastive Learningmentioning
confidence: 99%
“…Pinitas et al [61] proposed a model to learn general affectinfused multi-modal representations from audio, video, Electrocardiography (ECG), and Electrodermal Activity (EDA) modalities. The model was built upon the contrastive learning framework introduced in [58].…”
Section: Contrastive Learningmentioning
confidence: 99%
“…Zhang et al [59] used a Convolutional LSTM and a 1D-CNN to extract spatio-temporal facial and bio-sensing features, respectively. Recently, Pinitas et al [43] employed Supervised Contrastive Learning on audiovisual and physiological data to model arousal. Unlike the aforementioned studies, this paper presents preliminary findings regarding longterm player engagement prediction from in-game footage and game controller input using CNNs.…”
Section: Affect Modellingmentioning
confidence: 99%
“…As this study aims to model long-term engagement via players' multimodal signals, we consider the following data pre-processing method. We split each participant's session (video) into overlapping time windows [29,43] using a sliding step of 1.5 seconds and a window length of 10 seconds, corresponding to 22, 541 samples in the entire clean dataset. The sliding step and window length are essential hyperparameters since they influence, respectively, the size of the dataset and the information contained in each window.…”
Section: Data Pre-processingmentioning
confidence: 99%
See 1 more Smart Citation