2011 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA) 2011
DOI: 10.1109/aspaa.2011.6082322
|View full text |Cite
|
Sign up to set email alerts
|

Contour representations of sound

Abstract: Abstract-Many signals are naturally described by continuous contours in the time-frequency plane, but standard time-frequency methods disassociate continuous structures into isolated "atoms" of energy. Here we propose a method that represents any discrete time-series as a set of time-frequency contours. The edges of the contours are defined by fixed points of a generalized reassignment algorithm. These edges are linked together by continuity such that each contour represents a single phase-coherent region of t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2012
2012
2012
2012

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 20 publications
0
2
0
Order By: Relevance
“…First, a template vocalization of a selected syllable was manually chosen. Then, features from the template and the rest of the data were computed from a ratio of two quantities: the standard sonogram ( ) of the sound pressure time series , computed with a Gaussian window of time scale , and a sonogram computed with the derivative of the Gaussian window ( ) [29] . …”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…First, a template vocalization of a selected syllable was manually chosen. Then, features from the template and the rest of the data were computed from a ratio of two quantities: the standard sonogram ( ) of the sound pressure time series , computed with a Gaussian window of time scale , and a sonogram computed with the derivative of the Gaussian window ( ) [29] . …”
Section: Methodsmentioning
confidence: 99%
“…Using the ratio defined in (1), a group of sounds were aligned using their peak cross-correlation over the features mentioned above. A specified number of renditions were randomly chosen from the cluster of possible template matches to create an aggregate spectral density image, and a sparse time-frequency representation was calculated for each selected sound [29] . We denote the sparse time-frequency representation, , for sound i , and then apply the following transformation, …”
Section: Methodsmentioning
confidence: 99%