2023
DOI: 10.3390/electronics12040978
|View full text |Cite
|
Sign up to set email alerts
|

Music Emotion Recognition Based on a Neural Network with an Inception-GRU Residual Structure

Abstract: As a key field in music information retrieval, music emotion recognition is indeed a challenging task. To enhance the accuracy of music emotion classification and recognition, this paper uses the idea of inception structure to use different receptive fields to extract features of different dimensions and perform compression, expansion, and recompression operations to mine more effective features and connect the timing signals in the residual network to the GRU module to extract timing features. A one-dimension… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 28 publications
0
5
0
Order By: Relevance
“…The proposed neural network was compared with recently proposed models, as represented in Table 4, including two types of preprocessed datasets. The inception-GRU Residual Structure method [12] achieved 84.23 % accuracy in music emotion classification on the Soundtrack dataset. Improved deep belief network [13] achieved 83.35% accuracy.…”
Section: Experiments and Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…The proposed neural network was compared with recently proposed models, as represented in Table 4, including two types of preprocessed datasets. The inception-GRU Residual Structure method [12] achieved 84.23 % accuracy in music emotion classification on the Soundtrack dataset. Improved deep belief network [13] achieved 83.35% accuracy.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…Notably, a novel approach utilizing an Inception-GRU residual structure has been put forth, capturing the intricacies of musical expressions with significant efficacy. This methodology, grounded in the spectral matrix derived from logarithmic short-time Fourier transform, has showcased promising results on the Soundtrack dataset, achieving an accuracy surpassing traditional machine learning models [12]. In this paper, the researches presented an optimized structure of the Inception-V1 model which combines different convolution layers in parallel, and a deeper matrix is formed by concatenation the results processed by the convolution layers.…”
Section: Literature Overviewmentioning
confidence: 99%
“…In this paper, the Inception-BiGRU 28 module is built, as shown in Figure 4 . First, the logging data is reconstructed, and the two-dimensional data obtained by the reconstruction is used as the input of the network model.…”
Section: Methodsmentioning
confidence: 99%
“…Chaudhary et al [22] use three stacked CNN layers to learn emotional features from music spectrograms. Han et al [23] use the idea of an Inception Module to extract features of different dimensions. Also, they use one-dimensional residual CNNs and Gate Recurrent Unit (GRU) to extract timing features.…”
Section: Deep Learning In Mermentioning
confidence: 99%