2013
DOI: 10.3389/fpsyg.2013.00292
|View full text |Cite
|
Sign up to set email alerts
|

On the Acoustics of Emotion in Audio: What Speech, Music, and Sound have in Common

Abstract: Without doubt, there is emotional information in almost any kind of sound received by humans every day: be it the affective state of a person transmitted by means of speech; the emotion intended by a composer while writing a musical piece, or conveyed by a musician while performing it; or the affective state connected to an acoustic event occurring in the environment, in the soundtrack of a movie, or in a radio play. In the field of affective computing, there is currently some loosely connected research concer… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
158
1
9

Year Published

2015
2015
2022
2022

Publication Types

Select...
5
3

Relationship

2
6

Authors

Journals

citations
Cited by 188 publications
(169 citation statements)
references
References 32 publications
(37 reference statements)
1
158
1
9
Order By: Relevance
“…Another avenue for further development, especially when data for larger groups of singers and actors become available, is the use of machine learning to assess the degree of generalizability of the underlying acoustic profiles. A recent study by Weninger et al (2013) shows that algorithms trained on emotional music are quite successful on emotional speech and vice versa, suggesting that this may be a very viable approach. Weniger et al show that the effect generalizes, to some extent, even to environmental sounds.…”
Section: Resultsmentioning
confidence: 99%
“…Another avenue for further development, especially when data for larger groups of singers and actors become available, is the use of machine learning to assess the degree of generalizability of the underlying acoustic profiles. A recent study by Weninger et al (2013) shows that algorithms trained on emotional music are quite successful on emotional speech and vice versa, suggesting that this may be a very viable approach. Weniger et al show that the effect generalizes, to some extent, even to environmental sounds.…”
Section: Resultsmentioning
confidence: 99%
“…Previous findings in [18] suggest that the expression of emotions in speaking and singing voice are related. Further, [12] concludes that similar methods and acoustic features can be used to automatically classify emotions in speech, polyphonic music, as well as emotions perceived by listeners in or associated by them with other, general sounds. This suggests that the methods for speech emotion recognition can be transferred to singing emotion recognition.…”
Section: Related Workmentioning
confidence: 98%
“…[10][11][12]), although the fact that emotions are visible in acoustic properties of the voice has been frequently acknowledged [13,14]. In particular, in music, emotions play a major role and singers must be able to easily express a wide range of emotions.…”
Section: Related Workmentioning
confidence: 99%
“…In this work, we employ the openSMILE feature extractor [36] to obtain the ComParE acoustic feature set of 65 low-level descriptors (LLD) (4 energy-related, 55 spectral and 6 voicing-related), which has been successfully applied for automatic recognition of paralinguistic phenomena [37]. The 65 LLD used are summarized in Table 3 in [38].…”
Section: Featuresmentioning
confidence: 99%
“…The audio features extracted for each sequence of the CONFER Database are down-sampled to 25 Hz frequency to match the frame rate of the video stream. Similarly to [39,37] the audio features of each sequence are z-normalized (each feature component is normalized to mean=0 and standard deviation=1). Visual features.…”
Section: Featuresmentioning
confidence: 99%