Abstract:We are presenting results comparison of three artificial intelligence algorithms in a classification of time series derived from musical excerpts in this paper. Algorithms were chosen to represent different principles of classification – statistic approach, neural networks and competitive learning. The first algorithm is a classical k-Nearest neighbours algorithm, the second algorithm is Multilayer Perceptron (MPL), an example of artificial neural network and the third one is a Learning Vector Quantization (LV… Show more
“…This method can be successfully used for such tasks clarinet vs. talking or heavy vs. silence Magnatagatune tags discrimination. The accuracy of such tasks can be as high as 98% [2] in case of heavy vs. silence or 92% in case of clarinet vs. talking recordings. These datasets represent well-separable data, explaining good classification results.…”
Section: Discussionmentioning
confidence: 99%
“…It can be used to improve clustering results of SOM by using data labels [8]. For adjusting the codebook vector algorithm moves the codebook vector according to (2) …”
Section: Methodsmentioning
confidence: 99%
“…These recordings are in addition labeled by humans with tags evaluating many aspect of recordings like volume, instrumentation, genre etc. We used this database also in our former experiments [4], [3] and [2].…”
This paper describes classification of sound recordings based on their audio features. This is useful for querying large datasets, searching for recordings with some desired content. We use musical recordings as well as birdsongs recordings, which usually have rich structure and contain a lot of patterns suitable for classification. We present two different classification methods, one for musical recordings and one for birdsongs. These methods are compared and their differences are discussed. We use feature vectors that capture the audio content of recording as a whole piece and then classify these feature vectors using combination of the Self-organizing map and the Learning Vector Quantization, which represent a powerful algorithm using unlabeled as well as labeled data. In case of birdsongs we use feature vectors representing time frames of a recording.
“…This method can be successfully used for such tasks clarinet vs. talking or heavy vs. silence Magnatagatune tags discrimination. The accuracy of such tasks can be as high as 98% [2] in case of heavy vs. silence or 92% in case of clarinet vs. talking recordings. These datasets represent well-separable data, explaining good classification results.…”
Section: Discussionmentioning
confidence: 99%
“…It can be used to improve clustering results of SOM by using data labels [8]. For adjusting the codebook vector algorithm moves the codebook vector according to (2) …”
Section: Methodsmentioning
confidence: 99%
“…These recordings are in addition labeled by humans with tags evaluating many aspect of recordings like volume, instrumentation, genre etc. We used this database also in our former experiments [4], [3] and [2].…”
This paper describes classification of sound recordings based on their audio features. This is useful for querying large datasets, searching for recordings with some desired content. We use musical recordings as well as birdsongs recordings, which usually have rich structure and contain a lot of patterns suitable for classification. We present two different classification methods, one for musical recordings and one for birdsongs. These methods are compared and their differences are discussed. We use feature vectors that capture the audio content of recording as a whole piece and then classify these feature vectors using combination of the Self-organizing map and the Learning Vector Quantization, which represent a powerful algorithm using unlabeled as well as labeled data. In case of birdsongs we use feature vectors representing time frames of a recording.
“…We use algorithm called Dynamic Time Warping (Müller, 2007) commonly used for time series comparison in this paper. Some other time series exploration approaches can be found in (Fejfar, 2011) and (Fejfar, 2012). The data was acquired in our network laboratory simulating network traffi c by downloading fi les, streaming audio and video simultaneously.…”
FEJFAR JIŘÍ ŠŤASTNÝ JIŘÍ, POKORNÝ MARTIN, BALEJ JIŘÍ, ZACH PETR: Analysis of sound data streamed over the network. Acta Universitatis Agriculturae et Silviculturae Mendelianae Brunensis, 2013, LXI, No. 7, pp. 2105-2110
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.