Speech Prosody 2014 2014
DOI: 10.21437/speechprosody.2014-6
|View full text |Cite
|
Sign up to set email alerts
|

Towards Automatic Recognition of Attitudes: Prosodic Analysis of Video Blogs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(5 citation statements)
references
References 1 publication
0
5
0
Order By: Relevance
“…The results obtained using MFCC (228 features) are set as baseline in this study. Previous studies [12,11,10,15] do not evaluate the MFCC features for attitude recognition and use a large number of features (even more than number of instances [15]) which may result in over-fitting of machine learning models due to curse of dimensionality. However, in this study we used MFCC features for the classification task and also reduced the dimensionality of feature set using PCA as well as the proposed new method (AFT).…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…The results obtained using MFCC (228 features) are set as baseline in this study. Previous studies [12,11,10,15] do not evaluate the MFCC features for attitude recognition and use a large number of features (even more than number of instances [15]) which may result in over-fitting of machine learning models due to curse of dimensionality. However, in this study we used MFCC features for the classification task and also reduced the dimensionality of feature set using PCA as well as the proposed new method (AFT).…”
Section: Resultsmentioning
confidence: 99%
“…However, they did not perform fusion of features. In a different study [11], authors analyzed prosodic features of vlogger and found that these features (F0, voice quality and intensity) are correlated with a vlogger attitude, while in [12] they analyzed audio-visual features of vloggers for their attitude recognition. In all of the above studies, authors extracted the acoustic features using statistical functions (e.g.…”
Section: Introductionmentioning
confidence: 99%
“…They are affected by certain biases such as reputation and self-presentation [9] as an individual is likely to answer the questions in a way that maintains the image they wish to portray to others or that they determine to be more socially desirable [10]. Since previous psychological studies [11]- [14] frequently show that personality traits can be reflected by human nonverbal behaviours, most existing personality computing approaches aim to directly recognise apparent personality traits from the target subject's audio [15]- [17], visual [18]- [20] or audio-visual behaviours [21], [22]. There is evidence suggesting that an individual's response to certain situations largely depends on their personalities [23].…”
Section: Introductionmentioning
confidence: 99%
“…Before describing the followed approach, we provide a brief literature review on automatic personality trait recognition. In the past, various approaches have been used for recognizing apparent personality traits from different modalities such as audio [4,5], text [6][7][8] and visual information [9,10]. As in other recognition problems, multimodal systems are also investigated to improve robustness of prediction [11][12][13][14].…”
Section: Introduction and Related Workmentioning
confidence: 99%
“…In recent approaches to personality impressions classification, Support Vector Machines (SVM) [38] have been widely used [5,8,12,14]. Recently, a learning approach called Extreme Learning Machines (ELM) that is similar to SVMs but providing faster learning schemes has become popular [39].…”
Section: Introduction and Related Workmentioning
confidence: 99%