This tutorial provides a comprehensive overview of the methodological approach to collecting and analyzing auditory brainstem responses to complex sounds (cABRs). cABRs provide a window into how behaviorally relevant sounds such as speech and music are processed in the brain. Because temporal and spectral characteristics of sounds are preserved in this subcortical response, cABRs can be used to assess specific impairments and enhancements in auditory processing. Notably, subcortical function is neither passive nor hardwired but dynamically interacts with higher-level cognitive processes to refine how sounds are transcribed into neural code. This experience-dependent plasticity, which can occur on a number of time scales (e.g., life-long experience with speech or music, short-term auditory training, online auditory processing), helps shape sensory perception. Thus, by being an objective and non-invasive means for examining cognitive function and experience-dependent processes in sensory activity, cABRs have considerable utility in the study of populations where auditory function is of interest (e.g., auditory experts such as musicians, persons with hearing loss, auditory processing and language disorders). This tutorial is intended for clinicians and researchers seeking to integrate cABRs into their clinical and/or research programs.
The effects of music training in relation to brain plasticity have caused excitement, evident from the popularity of books on this topic among scientists and the general public. Neuroscience research has shown that music training leads to changes throughout the auditory system that prime musicians for listening challenges beyond music processing. This effect of music training suggests that, akin to physical exercise and its impact on body fitness, music is a resource that tones the brain for auditory fitness. Therefore, the role of music in shaping individual development deserves consideration.
Music and speech are very cognitively demanding auditory phenomena generally attributed to cortical rather than subcortical circuitry. We examined brainstem encoding of linguistic pitch and found that musicians show more robust and faithful encoding compared with nonmusicians. These results not only implicate a common subcortical manifestation for two presumed cortical functions, but also a possible reciprocity of corticofugal speech and music tuning, providing neurophysiological explanations for musicians' higher language-learning ability.Both music and spoken language involve the use of functionally and acoustically complex sound and are generally attributed to the neocortex [1][2][3][4] . Less is known about how long-term experience using these complex sounds shapes subcortical circuitry and the context specificity and reciprocity of this tuning 5 . By measuring the frequency following response (FFR), which presumably originates from the auditory brainstem (inferior colliculus) and encodes the energy of the stimulus fundamental frequency (f 0 ) with high fidelity 6 , previous work 7 has found increased linguistic pitch pattern encoding in Mandarin-speaking subjects relative to English-speaking subjects. These results reflect Mandarin-speaking subjects' long-term exposure to linguistic pitch patterns, as Mandarin Chinese, a tone language, uses pitch to signal word meaning (for example, /ma/ spoken with high or rising pitch patterns means 'mother' or 'numb', respectively). Moreover, similar to research on short-term perceptual learning 8 , these results can be viewed as context specific (that is, linguistic experiences, subserved by the cortex, enhance the encoding of linguistic information at the Correspondence should be addressed to P.C.M.W. (pwong@northwestern.edu). 5 These authors contributed equally to this work.Note: Supplementary information is available on the Nature Neuroscience website. COMPETING INTERESTS STATEMENTThe authors declare no competing financial interests. Author Manuscript brainstem). The nonspecificity of this long-term usage effect, though largely unknown, is both theoretically interesting and clinically and educationally relevant. Nonspecificity would suggest that either speech-or music-related experience can tune sensory encoding in the auditory brainstem via the corticofugal pathway. Notably, this tuning, whether speech-or music-induced, would enhance all relevant auditory functions (both speech and music) subserved by the rostral brainstem. HHS Public AccessWe measured FFR responses to linguistic pitch patterns at the rostral brainstem in ten amateur musicians and ten nonmusicians who had no previous exposure to a tone language (see Supplementary Table 1 online). Musicians (instrumentalists) had at least 6 years of continuous musical training (mean = 10.7 years) starting at or before the age of 12.Nonmusicians had nomore than3 years (mean = 1.2 years) at any time in their life. Informed written consent was obtained from all subjects. While watching a video, subjects listened...
Musical training is known to modify cortical organization. Here, we show that such modifications extend to subcortical sensory structures and generalize to processing of speech. Musicians had earlier and larger brainstem responses than nonmusician controls to both speech and music stimuli presented in auditory and audiovisual conditions, evident as early as 10 ms after acoustic onset. Phaselocking to stimulus periodicity, which likely underlies perception of pitch, was enhanced in musicians and strongly correlated with length of musical practice. In addition, viewing videos of speech (lip-reading) and music (instrument being played) enhanced temporal and frequency encoding in the auditory brainstem, particularly in musicians. These findings demonstrate practice-related changes in the early sensory encoding of auditory and audiovisual information.brainstem ͉ plasticity ͉ visual ͉ multisensory language
Musical experience appears to enhance the ability to hear speech in challenging listening environments. Large group differences were found for QuickSIN, and the results also suggest that this enhancement is derived in part from musicians' enhanced working memory and frequency discrimination. For HINT, in which performance was not linked to frequency discrimination ability and was only moderately linked to working memory, musicians still performed significantly better than the nonmusicians. The group differences for HINT were evident in the most difficult condition in which the speech and noise were presented from the same location and not spatially segregated. Understanding which cognitive and psychoacoustic factors as well as which lifelong experiences contribute to SIN may lead to more effective remediation programs for clinical populations for whom SIN poses a particular perceptual challenge. These results provide further evidence for musical training transferring to nonmusical domains and highlight the importance of taking musical training into consideration when evaluating a person's SIN ability in a clinical setting.
Musicians have lifelong experience parsing melodies from background harmonies, which can be considered a process analogous to speech perception in noise. To investigate the effect of musical experience on the neural representation of speech-in-noise, we compared subcortical neurophysiological responses to speech in quiet and noise in a group of highly trained musicians and nonmusician controls. Musicians were found to have a more robust subcortical representation of the acoustic stimulus in the presence of noise. Specifically, musicians demonstrated faster neural timing, enhanced representation of speech harmonics, and less degraded response morphology in noise. Neural measures were associated with better behavioral performance on the Hearing in Noise Test (HINT) for which musicians outperformed the nonmusician controls. These findings suggest that musical experience limits the negative effects of competing background noise, thereby providing the first biological evidence for musicians' perceptual advantage for speech-in-noise.
Older adults frequently report they can hear what is said but cannot understand the meaning, especially in noise. This difficulty may arise from the inability to process rapidly changing elements of speech. Aging is accompanied by a general slowing of neural processing and decreased neural inhibition, both of which likely interfere with temporal processing in auditory and other sensory domains. Age-related reductions in inhibitory neurotransmitter levels and delayed neural recovery can contribute to decreases in the auditory system’s temporal precision. Decreased precision may lead to neural timing delays, reductions in neural response magnitude, and a disadvantage in processing the rapid acoustic changes in speech. The auditory brainstem response (ABR), a scalp-recorded electrical potential, is known for its ability to capture precise neural synchrony within subcortical auditory nuclei; therefore, we hypothesized that a loss of temporal precision results in subcortical timing delays and decreases in response consistency and magnitude. To assess this hypothesis, we recorded ABRs to the speech syllable /da/ in normal hearing younger (ages 18 to 30) and older adult humans (60 to 67). Older adults had delayed ABRs, especially in response to the rapidly changing formant transition, and greater response variability. We also found that older adults had decreased phase locking and smaller response magnitudes than younger adults. Taken together, our results support the theory that older adults have a loss of temporal precision in subcortical encoding of sound, which may account, at least in part, for their difficulties with speech perception.
Children with learning problems often cannot discriminate rapid acoustic changes that occur in speech. In this study of normal children and children with learning problems, impaired behavioral discrimination of a rapid speech change (/dalpha/versus/galpha/) was correlated with diminished magnitude of an electrophysiologic measure that is not dependent on attention or a voluntary response. The ability of children with learning problems to discriminate another rapid speech change (/balpha/versus/walpha/) also was reflected in the neurophysiology. These results indicate that some children's discrimination deficits originate in the auditory pathway before conscious perception and have implications for differential diagnosis and targeted therapeutic strategies for children with learning disabilities and attention disorders.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
334 Leonard St
Brooklyn, NY 11211
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.