The efferent auditory nervous system may be a potent force in shaping how the brain responds to behaviorally significant sounds. Previous human experiments using the frequency following response (FFR) have shown efferent-induced modulation of subcortical auditory function online and over short- and long-term time scales; however, a contemporary understanding of FFR generation presents new questions about whether previous effects were constrained solely to the auditory subcortex. The present experiment used sine-wave speech (SWS), an acoustically-sparse stimulus in which dynamic pure tones represent speech formant contours, to evoke FFRSWS. Due to the higher stimulus frequencies used in SWS, this approach biased neural responses toward brainstem generators and allowed for three stimuli (/bɔ/, /bu/, and /bo/) to be used to evoke FFRSWSbefore and after listeners in a training group were made aware that they were hearing a degraded speech stimulus. All SWS stimuli were rapidly perceived as speech when presented with a SWS carrier phrase, and average token identification reached ceiling performance during a perceptual training phase. Compared to a control group which remained naïve throughout the experiment, training group FFRSWS amplitudes were enhanced post-training for each stimulus. Further, linear support vector machine classification of training group FFRSWS significantly improved post-training compared to the control group, indicating that training-induced neural enhancements were sufficient to bolster machine learning classification accuracy. These results suggest that the efferent auditory system may rapidly modulate auditory brainstem representation of sounds depending on their context and perception as non-speech or speech.
Bimodal hearing, which combines a cochlear implant (CI) with a contralateral hearing aid, provides significant speech recognition benefits in quiet and noise. These benefits have also been observed in normal-hearing listeners using vocoder-based CI simulation combined with low-pass filtered speech, even with acoustic bandwidths as narrow as 125–250 Hz. However, it is challenging to measure the optimal acoustic amplification with difficult-to-test populations, such as young children and adults with disabilities. The frequency following response (FFR) offers a potential solution to this problem, as it objectively quantifies subcortical phase-locking to speech features. Recently, FFR fundamental frequency amplitude in the non-implanted ear was found to be well-correlated with bimodal benefit in CI patients. The present study aimed to parametrically examine acoustic bandwidth effects (125, 250, 500, and 750 Hz) on speech evoked FFRs using simulated bimodal stimuli. We hypothesized that FFRenv amplitudes would increase as bandwidth increases up to 750 Hz and the minimal acoustic bandwidth needed to derive FFR bimodal benefit is less than 250 Hz.
Temporal acuity is the ability to differentiate between successive sounds based on temporal fluctuations in the waveform envelope. Psychophysically, human listeners can detect a gap as short as 2.5 ms between consecutive segments to encode the acoustical messages. The background noise diminishes the ability to follow fast variation between segments. In this study, we determined whether a physiological correlate of temporal acuity is also affected by the presence of noise. We recorded the auditory brainstem response (ABR) from human listeners using a harmonic complex followed by tone burst with the latter serving as the evoking stimulus. The duration and the depth of the silent gap between the harmonic complex and tone burst were manipulated. The latency of the ABR increased significantly as gap duration increased and gap depth decreased. No significant changes in amplitude were observed. These findings suggest that changing gap duration and depth affect the auditory system’s ability to encode successive sounds.
Vowel productions of 2 Mandarin-speaking children were audio recorded in their homes with picture naming tasks once every 3 months, from birth to 9 years old. The present study is the ninth year of a longitudinal observation. Major findings in this stage are: 1) The trend of decrease in formant values was continuously observed in the boy subject. As for the girl subject, it is not until nine years old, the obvious decrease in formant values was found, especially in F1; 2) F1 values are more stable than F2 values in both subjects. They appeared to acquire jaw movement sooner than tongue movement; 3) Throughout these 9 years, the variability of F1 is around 200-300Hz, and the variability of F2 is 500-700Hz in both subjects. No trend of decrease was found; 4) The trend of shrinkage in F1-F2 vowel area continues from 7 to 9 years old for the boy subject, but not for the girl subject; 5) There is a clear decline in fundamental frequencies at 8-9 years of age in the boy subject. Longitudinal data of vowel formant values from the same group of subjects provides important references for assessment and treatment of articulation disorders in children.
Speech frequency following responses (sFFRs) are increasingly used in translational auditory research. Statistically-based automated sFFR detection could aid response identification and provide a basis for stopping rules when recording responses in clinical and/or research applications. In this brief report, sFFRs were measured from 18 normal hearing adult listeners in quiet and speech-shaped noise. Two statistically-based automated response detection methods, the F-test and Hotelling’s T2 (HT2) test, were compared based on detection accuracy and test time. Similar detection accuracy across statistical tests and conditions was observed, although the HT2 test time was less variable. These findings suggest that automated sFFR detection is robust for responses recorded in quiet and speech-shaped noise using either the F-test or HT2 test. Future studies evaluating test performance with different stimuli and maskers are warranted to determine if the interchangeability of test performance extends to these conditions.
Temporal acuity is the ability to differentiate between sounds based on fluctuations in the waveform envelope. The proximity of successive sounds and background noise diminishes the ability to track rapid changes between consecutive sounds. We determined whether a physiological correlate of temporal acuity is also affected by these factors. We recorded the auditory brainstem response (ABR) from human listeners using a harmonic complex (S1) followed by a brief tone burst (S2) with the latter serving as the evoking signal. The duration and depth of the silent gap between S1 and S2 were manipulated, and the peak latency and amplitude of wave V were measured. The latency of the responses decreased significantly as the duration or depth of the gap increased. The amplitude of the responses was not affected by the duration or depth of the gap. These findings suggest that changing the physical parameters of the gap affects the auditory system’s ability to encode successive sounds.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.