Proceedings of the 2015 ACM on International Conference on Multimodal Interaction 2015
DOI: 10.1145/2818346.2820765
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating Speech, Face, Emotion and Body Movement Time-series Features for Automated Multimodal Presentation Scoring

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
17
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 40 publications
(17 citation statements)
references
References 13 publications
0
17
0
Order By: Relevance
“…Recent developments in multimodal data capture techniques have made it possible to detect these behavioral expressions during public speaking situations to assess the performance of participants. Ramanarayanan et al 21 extracted several nonverbal behavioral cues (head pose, gaze, and facial expressions) and speaking proficiency features (prosody, pronunciation and grammar) from 17 participants during five public speaking tasks. The authors correlated these extracted features with human-rated scores to find the modalities that are useful in predicting public speaking performance.…”
Section: Psychological Stress and Performance Assessment In Public Speaking Situationsmentioning
confidence: 99%
“…Recent developments in multimodal data capture techniques have made it possible to detect these behavioral expressions during public speaking situations to assess the performance of participants. Ramanarayanan et al 21 extracted several nonverbal behavioral cues (head pose, gaze, and facial expressions) and speaking proficiency features (prosody, pronunciation and grammar) from 17 participants during five public speaking tasks. The authors correlated these extracted features with human-rated scores to find the modalities that are useful in predicting public speaking performance.…”
Section: Psychological Stress and Performance Assessment In Public Speaking Situationsmentioning
confidence: 99%
“…We recently analyzed (Chen, Leong, Feng, Lee, & Somasundaran, 2015;Chen et al, 2014;Ramanarayanan, Chen, Leong, Feng, & Suendermann-Oeft, 2015) how fusing features obtained from different multimodal data streams, such as speech, face, body movement, and emotion tracks, can be applied to the scoring of multimodal presentations. We analyzed multimodal data collected by Chen et al (2015) consisting of synchronized and preprocessed recordings from 56 sessions involving speakers giving presentations on different topics.…”
Section: Multimodal Assessmentmentioning
confidence: 99%
“…We further observed that these features allowed us to achieve a prediction performance better than the human interrater agreement (as measured by the correlation between scores provided by two nonexpert human raters) for a subset of these scores. For further details, please see Ramanarayanan, Chen, et al ().…”
Section: Multimodal Assessmentmentioning
confidence: 99%
“…Emotions can also be measured through interaction with gesture, gaze and auditory stimuli. For instance, Ramanarayanan et al () used a variety of equipment and software tools including Microsoft Kinect to evaluate the quality of public presentations in relation to speech, face, emotion and body movement. Lu and Petiot () used a set of auditory stimuli to convey and assess a set of emotions such as funny, serious, relaxed, and depressed.…”
Section: Introductionmentioning
confidence: 99%