2013 Humaine Association Conference on Affective Computing and Intelligent Interaction 2013
DOI: 10.1109/acii.2013.102
|View full text |Cite
|
Sign up to set email alerts
|

User-centric Affective Video Tagging from MEG and Peripheral Physiological Responses

Abstract: Abstract-This paper presents a new multimodal database and the associated results for characterization of affect (valence, arousal and dominance) using the Magnetoencephalogram (MEG) brain signals and peripheral physiological signals (horizontal EOG, ECG, trapezius EMG). We attempt single-trial classification of affect in movie and music video clips employing emotional responses extracted from eighteen participants. The main findings of this study are that: (i) the MEG signal effectively encodes affective view… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2014
2014
2018
2018

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 20 publications
(19 citation statements)
references
References 22 publications
(43 reference statements)
0
19
0
Order By: Relevance
“…In this section, we attempt to predict the gold standard (or groundtruth) dynamic V/A ratings for each clip from audio-visual features using MTL, and show why learning the audio visual feature- (1) Average energy of audio signal [14] and first pitch frequency Formants (4) Formants up to 4400Hz Time frequency (8) mean and std of: MSpectrum flux, Spectral centroid, Delta spectrum magnitude, Band energy ratio [14] Zero crossing rate (1) Average zero crossing rate of audio signal [14] Silence ratio (2) Mean and std of proportion of silence in a time window [5,14] Video features Description Brightness (6) Mean of: Lighting key, shadow proportion, visual details, grayness, median of Lightness for frames, mean of median saturation for frames Color Features (41) Color variance, 20-bin histograms for hue and lightness in HSV space VisualExcitement (1) Features as defined in [29] Motion (1) Mean inter-frame motion [15] emotion relationship simultaneously for the 12 movie scenes is more effective than learning scene-specific models. Fig.…”
Section: Experiments and Resultsmentioning
confidence: 99%
See 4 more Smart Citations
“…In this section, we attempt to predict the gold standard (or groundtruth) dynamic V/A ratings for each clip from audio-visual features using MTL, and show why learning the audio visual feature- (1) Average energy of audio signal [14] and first pitch frequency Formants (4) Formants up to 4400Hz Time frequency (8) mean and std of: MSpectrum flux, Spectral centroid, Delta spectrum magnitude, Band energy ratio [14] Zero crossing rate (1) Average zero crossing rate of audio signal [14] Silence ratio (2) Mean and std of proportion of silence in a time window [5,14] Video features Description Brightness (6) Mean of: Lighting key, shadow proportion, visual details, grayness, median of Lightness for frames, mean of median saturation for frames Color Features (41) Color variance, 20-bin histograms for hue and lightness in HSV space VisualExcitement (1) Features as defined in [29] Motion (1) Mean inter-frame motion [15] emotion relationship simultaneously for the 12 movie scenes is more effective than learning scene-specific models. Fig.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…In a recent CS-based media tagging work, Soleymani et al [20] presented a dataset of 1000 songs for music emotion analysis, each annotated continuously over time by at least 10 users. Nevertheless, movies denote multimedia stimuli that best approximate the real world and movie clips have been found to be more effective for eliciting emotions in viewers as compared to music video clips in [1], and that is why we believe continuous emotion prediction with movie stimuli is important in the context of affective media representation and modeling.…”
Section: Crowdsourcing For Affective Media Taggingmentioning
confidence: 99%
See 3 more Smart Citations