2014
DOI: 10.3844/jcssp.2014.2584.2592
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Music Emotion Classification Using Artificial Neural Network Based on Vocal and Instrumental Sound Timbres

Abstract: Detecting emotion features in a song remains as a challenge in various area of research especially in Music Emotion Classification (MEC). In order to classify selected song with certain mood or emotion, the algorithms of the machine learning must be intelligent enough to learn the data features as to match the features accordingly to the accurate emotion. Until now, there were only few studies on MEC that exploit audio timbre features from vocal part of the song incorporated with the instrumental part of a son… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(5 citation statements)
references
References 14 publications
(15 reference statements)
0
5
0
Order By: Relevance
“…As a result, we are led to extract features related to the spectrum to enable the classification of diverse snoring sounds. Using the spectrum, eight features are calculated, namely, spectral centroid, spectral spread, spectral flatness, spectral decay point, spectral skewness, spectral slope, spectral entropy, and PR800 [20][21][22][23][24][25][26]. The first seven features are all derived from the spectrum obtained via the Fast Fourier Transform.…”
Section: Frequency-domain Feature Extractionmentioning
confidence: 99%
“…As a result, we are led to extract features related to the spectrum to enable the classification of diverse snoring sounds. Using the spectrum, eight features are calculated, namely, spectral centroid, spectral spread, spectral flatness, spectral decay point, spectral skewness, spectral slope, spectral entropy, and PR800 [20][21][22][23][24][25][26]. The first seven features are all derived from the spectrum obtained via the Fast Fourier Transform.…”
Section: Frequency-domain Feature Extractionmentioning
confidence: 99%
“…Some studies tried to evaluate the timbre difference by quantitative methods [75,76]. The major feature is a spectral centroid [77,78]. We calculated spectral centroids and 'microwave ding' was the most distant from the basic sounds (e.g.…”
Section: Natural Sound Selectionmentioning
confidence: 99%
“…In various research related to MIR, it used various kinds of data mining method for grouping including data classification and clustering such as C4.5 [4], decission tree [5], Support Vector Model [1], Artificial Neural Network [6], Self Organization [7], K-Means [8], [9], etc. Classification process by using this data mining algorithm was initiated with pre-processing stage.…”
Section: Introductionmentioning
confidence: 99%
“…In this stage music part to be used is the refrain, which the part with frequent words and notes repetition and this in the part that most determine mood type include in the music [10]. This part with 30 seconds duration [1], [6] with format of *.wav mono channel is furthermore processed by using signal processing of Fast Fourier Transform (FFT) and nine types…”
Section: Introductionmentioning
confidence: 99%