2009
DOI: 10.1162/comj.2009.33.4.23
|View full text |Cite
|
Sign up to set email alerts
|

Expressive Concatenative Synthesis by Reusing Samples from Real Performance Recordings

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0
4

Year Published

2010
2010
2023
2023

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 21 publications
(16 citation statements)
references
References 26 publications
0
12
0
4
Order By: Relevance
“…1 Note that the reconstructed audio examples sound rather unnatural because the experiment is not conducted in a full concatenative synthesis framework. In particular we use a uniform grain duration of 100ms and impose no temporal constraints, whereas a full concatenative synthesis system typically segments sounds using detected onsets and includes temporal constraints for continuity, and therefore is able to synthesise much more natural attack/sustain dynamics [Maestre et al, 2009]. Our method shows promise as the timbral component of a multi-attribute search which could potentially be used in concatenative synthesis, as well as other applications requiring timbral search from audio examples (e.g.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…1 Note that the reconstructed audio examples sound rather unnatural because the experiment is not conducted in a full concatenative synthesis framework. In particular we use a uniform grain duration of 100ms and impose no temporal constraints, whereas a full concatenative synthesis system typically segments sounds using detected onsets and includes temporal constraints for continuity, and therefore is able to synthesise much more natural attack/sustain dynamics [Maestre et al, 2009]. Our method shows promise as the timbral component of a multi-attribute search which could potentially be used in concatenative synthesis, as well as other applications requiring timbral search from audio examples (e.g.…”
Section: Resultsmentioning
confidence: 99%
“…Concatenative synthesisers typically operate not only on timbre, but use pitch and duration as well as temporal continuity constraints in their search strategy, and then modify the selected grains further to improve the match [Maestre et al, 2009]. While recognising the importance of these aspects in a full concatenative synthesis system, we designed an experiment in which the role of pitch, duration and temporal continuity were minimised, by excluding such factors from grain construction/analysis/resynthesis, and also by selecting audio excerpts whose variation is primarily timbral.…”
Section: Experiments 2: Concatenative Synthesismentioning
confidence: 99%
“…Given the set of source distributions used for obtaining the mixed distribution (see Section V-A), the cost is computed as a weighted sum of negative log-likelihoods of the vector , each one computed given the corresponding th original distribution. This is expressed as (38) where represents the weight applied to each of the likelihoods, and is an additional cost related to the bow displacement of the current execution candidate. The value of used for weighting each th likelihood is set to the Mahalanobis distance from the target performance context point to the centroid of the th source distribution (see Section V-A), computed as in (25).…”
Section: B Cost Computationmentioning
confidence: 99%
“…1) Sample Retrieval: As a first step, a sample candidate list is generated for each th note in the input score by making use of different sample annotations: matching articulation, bow direction, played string, bow context, and silence context (see Section III-A). Then, sample retrieval is performed through an optimal path search in a similar fashion as presented in [38] after [39] (again making use of dynamic programming). In this work, the main novelty is the introduction of a measure of distance between target bowing control parameters and sample bowing control parameters, therefore enabling the search for samples by accounting for bowing parameter contours.…”
Section: Sample-based Sound Synthesismentioning
confidence: 99%
“…More recently, Maestre et al used concatenative synthesis techniques aided by an expressive performance model to "generate an expressive audio sequence from the analysis of an arbitrary input score" [29]. Inspired by previous work by Schwarz, the authors gleaned segment information not only from analysis of the signal of an instrument note, but from the context of the score during the time that the note was played.…”
Section: Concatenative Synthesismentioning
confidence: 99%