2014 IEEE International Conference on Multimedia and Expo (ICME) 2014
DOI: 10.1109/icme.2014.6890290
|View full text |Cite
|
Sign up to set email alerts
|

Towards time-varying music auto-tagging based on CAL500 expansion

Abstract: Music auto-tagging refers to automatically assigning semantic labels (tags) such as genre, mood and instrument to music so as to facilitate text-based music retrieval. Although significant progress has been made in recent years, relatively little research has focused on semantic labels that are time-varying within a track. Existing approaches and datasets usually assume that different fragments of a track share the same tag labels, disregarding the tags that are time-varying (e.g., mood) or local in time (e.g.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 19 publications
(11 citation statements)
references
References 26 publications
0
11
0
Order By: Relevance
“…The CAL500 dataset includes 500 popular songs from Western countries with semantic labels derived from human listeners. One CNN-based music emotion classification [ 32 ] method for the CAL500 dataset, as well as its enriched version (CAL500exp) [ 33 ], was used for the classification of 18 emotion tags in the dataset. Recently, after music source separation [ 34 ] and attention [ 35 ], individual music sources were also applied to improve prediction of emotions in music by using a spectral representation of audio as the input.…”
Section: Related Workmentioning
confidence: 99%
“…The CAL500 dataset includes 500 popular songs from Western countries with semantic labels derived from human listeners. One CNN-based music emotion classification [ 32 ] method for the CAL500 dataset, as well as its enriched version (CAL500exp) [ 33 ], was used for the classification of 18 emotion tags in the dataset. Recently, after music source separation [ 34 ] and attention [ 35 ], individual music sources were also applied to improve prediction of emotions in music by using a spectral representation of audio as the input.…”
Section: Related Workmentioning
confidence: 99%
“…CAL500exp is an enriched version of the well-known CAL500. Wang et al [16] published the dataset. Labels of CAL500exp are annotated in the segment level instead of track level in CAL500.…”
Section: Datasetmentioning
confidence: 99%
“…Results on CAL500exp published. The first line of numerical parts shows the result of [16]. We list both of our results at the following two lines on CAL500exp.…”
Section: Model Structurementioning
confidence: 99%
See 2 more Smart Citations