2021
DOI: 10.3390/app112411926
|View full text |Cite
|
Sign up to set email alerts
|

A General Framework for Visualization of Sound Collections in Musical Interfaces

Abstract: While audio data play an increasingly central role in computer-based music production, interaction with large sound collections in most available music creation and production environments is very often still limited to scrolling long lists of file names. This paper describes a general framework for devising interactive applications based on the content-based visualization of sound collections. The proposed framework allows for a modular combination of different techniques for sound segmentation, analysis, and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 26 publications
0
5
0
Order By: Relevance
“…The original waveforms are firstly analyzed by the method of Mel-frequency cepstral coefficients (MFCCs), [38,39] and then processed by the feature extraction method to reduce the dimension of each sample while keeping most of the information. [40,41] Details of the pre-processing are shown in Supporting Information. [34] After pre-processing, each audio segment is represented by a two-dimension feature vector and is ready to be analyzed by quantum processor.…”
Section: Resultsmentioning
confidence: 99%
“…The original waveforms are firstly analyzed by the method of Mel-frequency cepstral coefficients (MFCCs), [38,39] and then processed by the feature extraction method to reduce the dimension of each sample while keeping most of the information. [40,41] Details of the pre-processing are shown in Supporting Information. [34] After pre-processing, each audio segment is represented by a two-dimension feature vector and is ready to be analyzed by quantum processor.…”
Section: Resultsmentioning
confidence: 99%
“…The value is then assumed to be a folder of audio samples. The samples are analyzed using the FluidCorpusMap library (Roma et al 2021). This library performs analysis and dimensionality reduction of the audio features and maps them to a grid using an assignment algorithm.…”
Section: Methodsmentioning
confidence: 99%
“…Corpus-based concatenative synthesis systems such as CataRT (Schwarz, Beller, Verbrugghe and Britton 2006) have traditionally been based on 2D scatterplots using scalar descriptors as axes. Several systems have explored dimensionality reduction and layout mapping for descriptor-based visualisation of sound collections in two dimensions (see Roma, Xambó, Green and Tremblay 2021, and references therein). These systems have generally been controlled through input devices such as mice, tablets, or other sensors.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations