2017
DOI: 10.3389/fpsyg.2017.00587
|View full text |Cite
|
Sign up to set email alerts
|

Perceptually Salient Regions of the Modulation Power Spectrum for Musical Instrument Identification

Abstract: The ability of a listener to recognize sound sources, and in particular musical instruments from the sounds they produce, raises the question of determining the acoustical information used to achieve such a task. It is now well known that the shapes of the temporal and spectral envelopes are crucial to the recognition of a musical instrument. More recently, Modulation Power Spectra (MPS) have been shown to be a representation that potentially explains the perception of musical instrument sounds. Nevertheless, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
2
1
1

Relationship

2
7

Authors

Journals

citations
Cited by 17 publications
(17 citation statements)
references
References 34 publications
0
17
0
Order By: Relevance
“…Future work will further verify the approach on other datasets with playing techniques, such as Studio-Online [6] and ConTimbre [18]. We will also compare the jTFST with other equivalent time-frequency representations, such as the two-dimensional Fourier transform and the modulation spectra [19].…”
Section: Resultsmentioning
confidence: 99%
“…Future work will further verify the approach on other datasets with playing techniques, such as Studio-Online [6] and ConTimbre [18]. We will also compare the jTFST with other equivalent time-frequency representations, such as the two-dimensional Fourier transform and the modulation spectra [19].…”
Section: Resultsmentioning
confidence: 99%
“…In the visual domain, these techniques have been extended in recent years to address not only low-level sensory processes, but higher-level cognitive mechanisms in humans: facial recognition [12], emotional expressions [2, 13], social traits [14], as well as their associated individual and cultural variations ([15]; for a review, see [5]). In speech, even more recently, reverse correlation and the associated “bubbles” technique were used to study spectro-temporal regions underlying speech intelligibility [16, 17] or phoneme discrimination in noise [18, 19] and, in music, timbre recognition of musical instruments [20, 21].…”
Section: Introductionmentioning
confidence: 99%
“…From a more general perspective, the current approach is in line with an upsurge of interest in signal analysis/re-synthesis approaches to the study of auditory perception (Mc-Dermott and Simoncelli, 2011;Overath et al, 2015;Ponsot et al, 2018;Thoret et al, 2017).…”
Section: Discussionmentioning
confidence: 86%
“…Most recently, Thoret et al (2016Thoret et al ( , 2017 showed that instrument identification is determined by specific instrument-specific spectrotemporal modulations, although their approach did not allow them to draw specific conclusions about the role of onsets. Ogg et al (2017) studied the minimal duration required to discriminate between musical instrument sounds, human speech, and human environmental sounds.…”
Section: A Previous Researchmentioning
confidence: 99%