2017
DOI: 10.1007/978-3-319-67220-5_3
|View full text |Cite
|
Sign up to set email alerts
|

Evaluation of Gated Recurrent Neural Networks in Music Classification Tasks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
7

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(8 citation statements)
references
References 10 publications
0
8
0
Order By: Relevance
“…The extracted features and segment representation of the initial frame were combined to obtain the fusional segment feature, which achieved an accuracy of 89.71% . Both GRU and LSTM were used to classify music and achieved 92% and 89% accuracy on the GTZAN dataset by Jukubik [28]. GRU and LSTM solve gradient vanishing and exploding problems of vanilla RNN, but they both are still susceptible to gradient decay as they both use sigmoid and hyperbolic tangent functions.…”
Section: Literature Reviewmentioning
confidence: 99%
See 1 more Smart Citation
“…The extracted features and segment representation of the initial frame were combined to obtain the fusional segment feature, which achieved an accuracy of 89.71% . Both GRU and LSTM were used to classify music and achieved 92% and 89% accuracy on the GTZAN dataset by Jukubik [28]. GRU and LSTM solve gradient vanishing and exploding problems of vanilla RNN, but they both are still susceptible to gradient decay as they both use sigmoid and hyperbolic tangent functions.…”
Section: Literature Reviewmentioning
confidence: 99%
“…A BN, ReLU activation, convolution, and average-pooling made the transition layer. In the final decision layer, global average pooling [28] took the average of each feature map to form a resulting vector and Copyright © 2021 MECS I.J. Information Technology and Computer Science, 2021, 2, 1-14 fed it to a softmax log-loss function, which produced a distribution over genre labels.…”
Section: Construction Of the Used Bbnnmentioning
confidence: 99%
“…The output of CNN is pooled to a feature map of a smaller dimension as an input to RNN, keeping an aptitude to process sequential data by utilizing its memory unit. By this ability, it examines the long term dependency and important patterns hidden in sequential data explained in [36]. A Global Layer Regularization (GLR) is applied to the combined model that reduces the dimensions by computing statistics across each feature and helps in finding optimal parameters quickly for training.…”
Section: The Proposed Globally Regularized Cnn-rnn Architecturementioning
confidence: 99%
“…Classical songs correctly predicted with overall 87% of data and Sufi songs correctly predicted with overall 82% of data. Jan Jakubik [7], they performed with two Recurrent Neural Network (RNN): LSTM and GRU. They experiment on 4 datasets: GTZAN, Emotify, Ballroom and LastFM.…”
Section: Related Workmentioning
confidence: 99%