Proceedings of the International Conference on Multimedia Information Retrieval 2010
DOI: 10.1145/1743384.1743400
|View full text |Cite
|
Sign up to set email alerts
|

Exploring automatic music annotation with "acoustically-objective" tags

Abstract: The task of automatically annotating music with text tags (referred to as autotagging) is vital to creating a large-scale semantic music discovery engine. Yet for an autotagging system to be successful, a large and cleanly-annotated data set must exist to train the system. For this reason, we have collected a data set, called Swat10k, which consists of 10,870 songs annotated using a vocabulary of 475 acoustic tags and 153 genre tags from Pandora's Music Genome Project. The acoustic tags are considered "acousti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
52
0

Year Published

2011
2011
2016
2016

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 62 publications
(52 citation statements)
references
References 15 publications
0
52
0
Order By: Relevance
“…2 represents segmentation process from IVS. Echo Nest timbre, Echo Nest pitches (ENP), and Echo Nest loudness (ENL) are computed using Echo Nest API [15]. Segments are characterized by their perceptual onsets of an Indian song in Echo Nest API automatically.…”
Section: Segmentationmentioning
confidence: 99%
See 1 more Smart Citation
“…2 represents segmentation process from IVS. Echo Nest timbre, Echo Nest pitches (ENP), and Echo Nest loudness (ENL) are computed using Echo Nest API [15]. Segments are characterized by their perceptual onsets of an Indian song in Echo Nest API automatically.…”
Section: Segmentationmentioning
confidence: 99%
“…Music researchers have started a company named the Echo Nest [15] in 2005. It gives free analysis of music via API.…”
Section: Feature Extractionmentioning
confidence: 99%
“…However, when the metadata cannot be used if the user doesn't know the title, this retrieval method cannot be applied. Therefore, there are studies related to the annotation [1][2] and the tagging [3], [4] in order to retrieve easily by enhancing the metadata. Moreover, there are researches which target music is retrieved by humming a part of the tune without using the metadata [5].…”
Section: Introductionmentioning
confidence: 99%
“…This is, in part, due to the time required to listen to recorded audio, contrasted to the time taken to interpret an image. In addition, automated techniques are not as advanced, and building a repository of annotations created by experts is not scalable and very expensive (Tingle, Kim, & Turnbull, 2010). There are relatively well-known attributes for images and text -for an image the dimensions, colour, composition style, and content; text is composed of groups of symbols with a recognisable font or style, which can be further grouped into words, sentences, and paragraphs.…”
Section: Acoustic Data Analysismentioning
confidence: 99%
“…Producing high-quality annotations can be done by employing birders dedicated to the task; however, this can be difficult to expand, due to financial restrictions or the time and effort required (Tingle et al, 2010).…”
Section: Acoustic Data Analysismentioning
confidence: 99%