2015
DOI: 10.1186/s13636-015-0057-6
|View full text |Cite
|
Sign up to set email alerts
|

Emotion in the singing voice—a deeperlook at acoustic features in the light ofautomatic classification

Abstract: We investigate the automatic recognition of emotions in the singing voice and study the worth and role of a variety of relevant acoustic parameters. The data set contains phrases and vocalises sung by eight renowned professional opera singers in ten different emotions and a neutral state. The states are mapped to ternary arousal and valence labels. We propose a small set of relevant acoustic features basing on our previous findings on the same data and compare it with a large-scale state-of-the-art feature set… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
18
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 26 publications
(19 citation statements)
references
References 19 publications
1
18
0
Order By: Relevance
“…Previous research suggests that the expression of emotions in the singing and speaking voice are related, 28 and that the same methods of emotion recognition apply to both. 29 As with the speaking voice, the voice quality in anger is easiest to recognize. This phenomenon might have an evolutionary underpinning, as it continues to be a useful skill to recognize potentially hazardous situations.…”
Section: Discussionmentioning
confidence: 99%
“…Previous research suggests that the expression of emotions in the singing and speaking voice are related, 28 and that the same methods of emotion recognition apply to both. 29 As with the speaking voice, the voice quality in anger is easiest to recognize. This phenomenon might have an evolutionary underpinning, as it continues to be a useful skill to recognize potentially hazardous situations.…”
Section: Discussionmentioning
confidence: 99%
“…In the ML experiments, we employ the ComParE feature set (Schuller et al 2013), comprising 6373 acoustic features (Eyben et al 2015) computed by applying statistical functions to 65 Low-Level Descriptors (LLDs), extracted by the OPENSMILE feature extractor (Eyben et al 2010), and a Support Vector Machine (SVM) classifier with a linear kernel from the open-source toolkit LIBLINEAR (Fan et al 2008). Even though Deep Neural Networks (DNNs) are prevalent nowadays for ML tasks, in affective computing research their performance is not yet superior to rather classic ML procedures such as SVMs.…”
Section: Methodsmentioning
confidence: 99%
“…Acoustic Features. Acoustic measures of music have been extensively studied (Berenzweig et al 2004;Mckay and Fujinaga 2008;Eyben et al 2015). In this work, we employ openSMILE (Eyben et al 2013), an open-source multimedia feature extractor to extract the "emobase" set of 988-dimensional acoustic features.…”
Section: Music Feature Extractionmentioning
confidence: 99%