Proceedings of the 22nd ACM International Conference on Multimedia 2014
DOI: 10.1145/2647868.2654924
|View full text |Cite
|
Sign up to set email alerts
|

An Event Driven Fusion Approach for Enjoyment Recognition in Real-time

Abstract: Social signals and interpretation of carried information is of high importance in Human Computer Interaction. Often used for affect recognition, the cues within these signals are displayed in various modalities. Fusion of multi-modal signals is a natural and interesting way to improve automatic classification of emotions transported in social signals. Throughout most present studies, uni-modal affect recognition as well as multi-modal fusion, decisions are forced for fixed annotation segments across all modali… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
29
0

Year Published

2015
2015
2020
2020

Publication Types

Select...
3
2
1

Relationship

4
2

Authors

Journals

citations
Cited by 27 publications
(29 citation statements)
references
References 28 publications
0
29
0
Order By: Relevance
“…Extracted facial and paralinguistic cues are combined through fusion strategies in order to generate a final prediction. Our work on fusion draws on Lingenfelser's [20] "event-driven" fusion, which is based on [13]. The algorithm does not force decisions throughout considered modalities for each time frame, but instead asynchronously fuses time-sensitive events from any given number of modi.…”
Section: Multimodal Communication Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…Extracted facial and paralinguistic cues are combined through fusion strategies in order to generate a final prediction. Our work on fusion draws on Lingenfelser's [20] "event-driven" fusion, which is based on [13]. The algorithm does not force decisions throughout considered modalities for each time frame, but instead asynchronously fuses time-sensitive events from any given number of modi.…”
Section: Multimodal Communication Analysismentioning
confidence: 99%
“…This has the advantage of incorporating temporal alignments between modi and being very flexible with respect to the type and mode of used events. In [20], this algorithm was used to combine the recognition of short-timed laugh (audio) and smile (video) events for a continuous assessment of a user's level of positive valence. For KRISTINA, it is extended to cover the whole valence arousal space, spanned by positive and negative valence and arousal axes.…”
Section: Multimodal Communication Analysismentioning
confidence: 99%
“…In [63], an event driven real-time fusion system was proposed. It rather corresponds to a late fusion mechanism, with some additional time-based accumulation.…”
Section: Multimodal Fusionmentioning
confidence: 99%
“…This decision determines if the observed window is labeled as a laughter window. A detailed description of the recognition system is found in [23]. Once trained on a subset of the corpus, it can be used to generate automated annotations for the remaining sessions.…”
Section: Automated Laughter Annotationsmentioning
confidence: 99%