2019
DOI: 10.1016/j.neuroimage.2019.01.075
|View full text |Cite
|
Sign up to set email alerts
|

Separable neural representations of sound sources: Speaker identity and musical timbre

Abstract: Human listeners can quickly and easily recognize different sound sources (objects and events) in their environment. Understanding how this impressive ability is accomplished can improve signal processing and machine intelligence applications along with assistive listening technologies. However, it is not clear how the brain represents the many sounds that humans can recognize (such as speech and music) at the level of individual sources, categories and acoustic features. To examine the cortical organization of… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
13
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
1
1

Relationship

2
6

Authors

Journals

citations
Cited by 17 publications
(13 citation statements)
references
References 54 publications
0
13
0
Order By: Relevance
“…While these acoustic qualities have previously been shown to influence the strength of neural responses on different time scales (Caclin et al, 2006;Giard et al, 1995;Rosburg, 2003;Näätänen & Picton, 1987;Roberts et al, 2000;Alluri et al, 2012;Giordano et al, 2012;Allen et al, 2017;Lewis et al, 2012;Warren et al, 2005;Lewis et al, 2009;Patterson et al, 2002;Menon et al, 2002;Langers et al, 2007;Norman-Haignere et al, 2015;Schönwiesner & Zatorre, 2009;Hullett et al, 2016), our results go further by beginning to link these features to representations of auditory objects and events early in perception. Additionally, we show that these features, which have typically been examined in fMRI (Alluri et al, 2012;Giordano et al, 2012;Allen et al, 2017;Lewis et al, 2012;Warren et al, 2005;Lewis et al, 2009;Patterson et al, 2002;Menon et al, 2002;Langers et al, 2007;Ogg et al, 2019;Norman-Haignere et al, 2015;Schönwiesner & Zatorre, 2009), vary in their influence over time, with more primary acoustic processes in cortex associated with decoding accuracy early in the response such as spectral envelope and modulation power spectra (Norman-Haignere et al, 2015, Chi et al, 2005, Theunissen & Elie, 2014. Later in the response, however, neural representations were more related to spectral and temporal centroid features (with spectral variability and aperiodicity correlating strongly throughout).…”
Section: Discussionmentioning
confidence: 84%
See 2 more Smart Citations
“…While these acoustic qualities have previously been shown to influence the strength of neural responses on different time scales (Caclin et al, 2006;Giard et al, 1995;Rosburg, 2003;Näätänen & Picton, 1987;Roberts et al, 2000;Alluri et al, 2012;Giordano et al, 2012;Allen et al, 2017;Lewis et al, 2012;Warren et al, 2005;Lewis et al, 2009;Patterson et al, 2002;Menon et al, 2002;Langers et al, 2007;Norman-Haignere et al, 2015;Schönwiesner & Zatorre, 2009;Hullett et al, 2016), our results go further by beginning to link these features to representations of auditory objects and events early in perception. Additionally, we show that these features, which have typically been examined in fMRI (Alluri et al, 2012;Giordano et al, 2012;Allen et al, 2017;Lewis et al, 2012;Warren et al, 2005;Lewis et al, 2009;Patterson et al, 2002;Menon et al, 2002;Langers et al, 2007;Ogg et al, 2019;Norman-Haignere et al, 2015;Schönwiesner & Zatorre, 2009), vary in their influence over time, with more primary acoustic processes in cortex associated with decoding accuracy early in the response such as spectral envelope and modulation power spectra (Norman-Haignere et al, 2015, Chi et al, 2005, Theunissen & Elie, 2014. Later in the response, however, neural representations were more related to spectral and temporal centroid features (with spectral variability and aperiodicity correlating strongly throughout).…”
Section: Discussionmentioning
confidence: 84%
“…And while musical training has sometimes been found to influence auditory cortical responses (Bidelman, Weiss, Moreno, & Alain, 2014), these effects might arise more robustly during demanding behavioral tasks (see also Alho et al, 2016;Bidelman & Walker, 2017). Indeed, other findings based on passive tasks with diverse, readily distinguishable stimuli have not found an influence of musical training on neural representations (Ogg et al, 2019).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…This supports our expectation that the responses 329 are temporally individualized for each sound. Additionally, it suggests that EEG does not pick 330 up the subtle spatial differences in cortical responsiveness to speech and music sounds, as has 331 been found in prior fMRI literature (Giordano et al, 2014;Norman-Haignere et al, 2015, 2019332 14 Ogg et al, 2019), which might be observed with many more speech and music samples in a 333 future experiment. 334…”
mentioning
confidence: 84%
“…To address the issue of time demands, we relied on short duration sounds (250-ms, including the sounds’ onsets). While this may appear to be a quite restricted duration, findings from gating paradigms demonstrate that durations over 200-ms support asymptotic performance for many sound identification tasks (Gray, 1942; Robinson and Patterson, 1995a,b; Suied et al, 2014; Ogg et al, 2017), and this sort of controlled, limited duration has commonly been used in many neuroimaging studies (e.g., Formisano et al, 2008; Leaver and Rauschecker, 2010; Ogg et al, 2019). We focused on sound onsets because they have been shown to carry a large amount of information important for sound perception (Iverson and Krumhansl, 1993; Hjortkjær and McAdams, 2016) and identification (Saldanha and Corso, 1964; Wedin and Goude, 1972; Lemaitre and Heller, 2012; Ogg et al, 2019) and this choice naturally focuses our investigation on the critical timescales when real-world identification is taking place (i.e., after a sound has been initiated in the environment).…”
Section: Introductionmentioning
confidence: 99%