Proceedings of the 1st International Workshop on Digital Libraries for Musicology 2014
DOI: 10.1145/2660168.2660176
|View full text |Cite
|
Sign up to set email alerts
|

Incremental Dataset Definition for Large Scale Musicological Research

Abstract: Conducting experiments on large scale musical datasets often requires the definition of a dataset as a first step in the analysis process. This is a classification task, but metadata providing the relevant information is not always available or reliable and manual annotation can be prohibitively expensive. In this study we aim to automate the annotation process using a machine learning approach for classification. We evaluate the effectiveness and the trade-off between accuracy and required number of annotated… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 19 publications
0
2
0
Order By: Relevance
“…The statistics of these datasets are listed in Table 1. The OpenBMAT [6] and ORF TV dataset [44] are collected from TV programs, and AVASpeech [10] comprises audio from YouTube, while Muspeak [14] has a variety of content such as concert, radio broadcast, and low-fidelity folk music.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The statistics of these datasets are listed in Table 1. The OpenBMAT [6] and ORF TV dataset [44] are collected from TV programs, and AVASpeech [10] comprises audio from YouTube, while Muspeak [14] has a variety of content such as concert, radio broadcast, and low-fidelity folk music.…”
Section: Methodsmentioning
confidence: 99%
“…Note that % of music/speech is estimated based on the duration labeled as music or speech and the total duration of the audio content means they can be used only for either speech or music detection, but not both. The GTZAN Speech and Music dataset [11], Scheirer & Slaney Music Speech [12], MUSAN [13], and Muspeak [14] datasets contain only short segments and non-overlapping speech or music labels. Thus, these datasets can only be used for a simplified music and speech segmentation task, where the audio segments can only be classified into either speech, music, or noise without any overlap.…”
Section: Introductionmentioning
confidence: 99%
“…-Adaptive learning: adaptive learning systems can continuously collect and interpret student data, change the direction and environment of students' learning, taking into account their needs and abilities [16].…”
Section: Big Data and Learning Analytics In General Education Schoolsmentioning
confidence: 99%