2013
DOI: 10.1109/mis.2013.34
|View full text |Cite
|
Sign up to set email alerts
|

YouTube Movie Reviews: Sentiment Analysis in an Audio-Visual Context

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
156
0
4

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 317 publications
(169 citation statements)
references
References 6 publications
0
156
0
4
Order By: Relevance
“…While recent work has started to study online video, it has been either in the passive viewer case as in [38] (that analyzed observers of video advertising who essentially do not talk), or has used limited facial expression cues (smiles only) in the context of online video reviews (not addressing the personality inference task) [52] [40]. In contrast to these works, our work studies a much richer set of facial expression cues derived from all the basic facial expressions as estimated by a FACs-based recognizer.…”
Section: Analyzing Personality Impressionsmentioning
confidence: 99%
“…While recent work has started to study online video, it has been either in the passive viewer case as in [38] (that analyzed observers of video advertising who essentially do not talk), or has used limited facial expression cues (smiles only) in the context of online video reviews (not addressing the personality inference task) [52] [40]. In contrast to these works, our work studies a much richer set of facial expression cues derived from all the basic facial expressions as estimated by a FACs-based recognizer.…”
Section: Analyzing Personality Impressionsmentioning
confidence: 99%
“…As future work there are several avenues likely to improve on our results. Beyond late fusion, other ways to combine prosodic and lexical similarity should be tried (Wollmer et al, 2013;Bruni et al, 2014). For example, recent developments in vector space representations of words (Turian et al, 2010;Erk, 2012;Mikolov et al, 2013;Huang et al, 2013), suggest that it could be productive to build a unified lexico-prosodic vector-space model of both meaning and dialog activity.…”
Section: Discussionmentioning
confidence: 99%
“…Most research relating to using prosody for audio search has focused on detecting dialog activities that people might like to search for. Prosody-based classifiers can, for example, spot interactional "hotspots" where the speakers are unusually involved (Wrede and Shriberg, 2003;Oertel et al, 2011), conflicts (Kim et al, 2012), agreements on action items (Purver et al, 2007), various emotional and attitudinal states and stances (Toivanen and Seppänen, 2002;Wollmer et al, 2013), and dialog acts such as question, apology, promise, and persuasion attempt (Larson et al, 2011;Freedman et al, 2011). This work has shown many dialog activities are indeed associated with characteristic prosodic features and patterns.…”
Section: Background: Prosody For Search In Speechmentioning
confidence: 99%
“…CNN trained on faces) features; for instance, recurrent neural network (RNN) and 3D convolutional networks (C3D), speci cally trained on faces, have been combined with audio features by Fan et al [12]. Wöllmer et al [33], try to understand the speaker's sentiment in on-line videos containing movie reviews by leveraging acoustic, visual and linguistic features. Rosas et al [24] use a similar approach to classify the speaker's emotion in Spanish videos.…”
Section: Related Workmentioning
confidence: 99%
“…Several works have proposed features for recognizing emotion in faces (e.g. [7,33]), but optimal features for this task are still unclear.…”
Section: Introductionmentioning
confidence: 99%