2016
DOI: 10.1016/j.neucom.2016.01.099
|View full text |Cite
|
Sign up to set email alerts
|

Bridge the semantic gap between pop music acoustic feature and emotion: Build an interpretable model

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 16 publications
(5 citation statements)
references
References 46 publications
0
5
0
Order By: Relevance
“…15 A model for music emotion recognition using feature selection and statistical models has been proposed. 16 The results obtained had a higher average accuracy rate for arousal compared to valence (80% for arousal vs. 63% for valence). The combination of audio, lyrics, and linguistic data to classify Greek songs into several valence and arousal categories has been introduced.…”
Section: Related Workmentioning
confidence: 86%
“…15 A model for music emotion recognition using feature selection and statistical models has been proposed. 16 The results obtained had a higher average accuracy rate for arousal compared to valence (80% for arousal vs. 63% for valence). The combination of audio, lyrics, and linguistic data to classify Greek songs into several valence and arousal categories has been introduced.…”
Section: Related Workmentioning
confidence: 86%
“…It is difficult to induce an individual’s emotions effectively and measure the level of emotional arousal accurately [ 40 ]. The methods most used in emotion induction research were movie clips [ 41 ], personalized recall [ 42 ], picture viewing [ 43 ], acoustic material [ 44 ] and standardized imagery [ 45 ]. Analysis results showed that music, movies and imagery were relatively ideal methods for emotion induction and the success rate was more than 75% [ 46 , 47 ].…”
Section: Study 2-driving Intention Prediction Models Adapting To Mmentioning
confidence: 99%
“…Many current methods for obtaining insights into deep network-based audio classification systems do not explain the predictions in a human understandable way but rather design special filters that can be visualized [13], or analyze neuron activations [8]. To the best of our knowledge, [19] is the only attempt to build an interpretable model for MER. They performed the task of feature extraction and selection and built models from different model classes on top of them.…”
Section: Related Workmentioning
confidence: 99%