2017
DOI: 10.48550/arxiv.1704.05665
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

CNN based music emotion classification

Xin Liu,
Qingcai Chen,
Xiangping Wu
et al.

Abstract: Music emotion recognition (MER) is usually regarded as a multi-label tagging task, and each segment of music can inspire specific emotion tags. Most researchers extract acoustic features from music and explore the relations between these features and their corresponding emotion tags. Considering the inconsistency of emotions inspired by the same music segment for human beings, seeking for the key acoustic features that really affect on emotions is really a challenging task. In this paper, we propose a novel ME… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 8 publications
0
5
0
Order By: Relevance
“…In [99], a CNN is developed for emotion classification with 18 emotion tags, using time and frequency domain information. The experiments make use of the CAL500 (Turnbull 2007 forward) and CAL500exp ( [100]) datasets.…”
Section: Convolutional Neural Network (Cnn)mentioning
confidence: 99%
“…In [99], a CNN is developed for emotion classification with 18 emotion tags, using time and frequency domain information. The experiments make use of the CAL500 (Turnbull 2007 forward) and CAL500exp ( [100]) datasets.…”
Section: Convolutional Neural Network (Cnn)mentioning
confidence: 99%
“…Different weights are allocated on different time blocks (chunks), and the song-level emotion classification prediction is obtained through fusion. Liu et al [22] regards music emotion recognition as a multi-label classification task, and uses convolutional neural networks and spectrum diagram to complete end-to-end classification. Chen et al [2] considered the complementarity between CNN with different structures and between CNN and LSTM, and combined multi-channel CNN with different structures and LSTM into a unified structure (Multi-channel Convolutional LSTM, MCCLSTM) to extract advanced music descriptors.…”
Section: Mer With Acoustic-onlymentioning
confidence: 99%
“…By converting the original data into a spectral graph, and then inputting the spectral graph into the CNN for emotion recognition. Liu and others [6] use the spectral graph computed by the short-time Fourier transform of the audio signal as input. Each music's spectral graph undergoes convolutional layers, pooling layers, and hidden layers, and finally, it goes through SoftMax for prediction.…”
Section: Introductionmentioning
confidence: 99%