2011
DOI: 10.1007/978-3-642-27169-4_9
|View full text |Cite
|
Sign up to set email alerts
|

A Comparison of Human, Automatic and Collaborative Music Genre Classification and User Centric Evaluation of Genre Classification Systems

Abstract: Abstract. In this paper two sets of evaluation experiments are conducted. First, we compare state-of-the-art automatic music genre classification algorithms to human performance on the same dataset, via a listening experiment. This will show that the improvements of contentbased systems over the last years have reduced the gap between automatic and human classification performance, but could not yet close this gap. As an important extension to previous work in this context, we will also compare the automatic a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
15
0

Year Published

2013
2013
2019
2019

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 11 publications
(16 citation statements)
references
References 6 publications
1
15
0
Order By: Relevance
“…The work proposing AdaBFFs (Bergstra et al 2006a), SRCAM (Panagakis et al 2009b), and the features of MAPsCAT (Andén and Mallat 2011), present only classification accuracy. Furthermore, based on classification accuracy, Seyerlehner et al (2010) argue that the performance gap between MGR systems and humans is narrowing; and in this issue, Humphrey et al conclude "progress in content-based music informatics is plateauing" (Humphrey et al 2013). Figure 2 shows that with respect to the classification accuracies in GTZAN reported in 83 published works (Sturm 2013b), those of AdaBFFs, SRCAM, and MAPsCAT lie above what is reported best in half of this work.…”
Section: Evaluating the Performance Statistics Of Mgr Systemsmentioning
confidence: 64%
See 4 more Smart Citations
“…The work proposing AdaBFFs (Bergstra et al 2006a), SRCAM (Panagakis et al 2009b), and the features of MAPsCAT (Andén and Mallat 2011), present only classification accuracy. Furthermore, based on classification accuracy, Seyerlehner et al (2010) argue that the performance gap between MGR systems and humans is narrowing; and in this issue, Humphrey et al conclude "progress in content-based music informatics is plateauing" (Humphrey et al 2013). Figure 2 shows that with respect to the classification accuracies in GTZAN reported in 83 published works (Sturm 2013b), those of AdaBFFs, SRCAM, and MAPsCAT lie above what is reported best in half of this work.…”
Section: Evaluating the Performance Statistics Of Mgr Systemsmentioning
confidence: 64%
“…The work by Vatolkin (2012) provides a comparison of various performance statistics for music classification. Other works (Berenzweig et al 2004;Craft et al 2007;Craft 2007;Lippens et al 2004;Wiggins 2009;Seyerlehner et al 2010;Sturm 2012b) argue for measuring performance in ways that take into account the natural ambiguity of music genre and similarity. For instance, we Sturm (2012b), Craft et al (2007) and Craft (2007) argue for richer experimental designs than having a system apply a single label to music with a possibly problematic "ground truth."…”
Section: Evaluation In Music Genre Recognition Researchmentioning
confidence: 99%
See 3 more Smart Citations