Speech recognition was measured as a function of spectral resolution (number of spectral channels) and speech-to-noise ratio in normal-hearing (NH) and cochlear-implant (CI) listeners. Vowel, consonant, word, and sentence recognition were measured in five normal-hearing listeners, ten listeners with the Nucleus-22 cochlear implant, and nine listeners with the Advanced Bionics Clarion cochlear implant. Recognition was measured as a function of the number of spectral channels (noise bands or electrodes) at signal-to-noise ratios of + 15, + 10, +5, 0 dB, and in quiet. Performance with three different speech processing strategies (SPEAK, CIS, and SAS) was similar across all conditions, and improved as the number of electrodes increased (up to seven or eight) for all conditions. For all noise levels, vowel and consonant recognition with the SPEAK speech processor did not improve with more than seven electrodes, while for normal-hearing listeners, performance continued to increase up to at least 20 channels. Speech recognition on more difficult speech materials (word and sentence recognition) showed a marginally significant increase in Nucleus-22 listeners from seven to ten electrodes. The average implant score on all processing strategies was poorer than scores of NH listeners with similar processing. However, the best CI scores were similar to the normal-hearing scores for that condition (up to seven channels). CI listeners with the highest performance level increased in performance as the number of electrodes increased up to seven, while CI listeners with low levels of speech recognition did not increase in performance as the number of electrodes was increased beyond four. These results quantify the effect of number of spectral channels on speech recognition in noise and demonstrate that most CI subjects are not able to fully utilize the spectral information provided by the number of electrodes used in their implant.
Objective: To update a 15-year-old study of 800 postlinguistically deaf adult patients showing how duration of severe to profound hearing loss, age at cochlear implantation (CI), age at onset of severe to profound hearing loss, etiology and CI experience affected CI outcome. Study Design: Retrospective multicenter study. Methods: Data from 2251 adult patients implanted since 2003 in 15 international centers were collected and speech scores in quiet were converted to percentile ranks to remove differences between centers. Results: The negative effect of long duration of severe to profound hearing loss was less important in the new data than in 1996; the effects of age at CI and age at onset of severe to profound hearing loss were delayed until older ages; etiology had a smaller effect, and the effect of CI experience was greater with a steeper learning curve. Patients with longer durations of severe to profound hearing loss were less likely to improve with CI experience than patients with shorter duration of severe to profound hearing loss. Conclusions: The factors that were relevant in 1996 were still relevant in 2011, although their relative importance had changed. Relaxed patient selection criteria, improved clinical management of hearing loss, modifications of surgical practice, and improved devices may explain the differences.
ObjectiveTo test the influence of multiple factors on cochlear implant (CI) speech performance in quiet and in noise for postlinguistically deaf adults, and to design a model of predicted auditory performance with a CI as a function of the significant factors.Study DesignRetrospective multi-centre study.MethodsData from 2251 patients implanted since 2003 in 15 international centres were collected. Speech scores in quiet and in noise were converted into percentile ranks to remove differences between centres. The influence of 15 pre-, per- and postoperative factors, such as the duration of moderate hearing loss (mHL), the surgical approach (cochleostomy or round window approach), the angle of insertion, the percentage of active electrodes, and the brand of device were tested. The usual factors, duration of profound HL (pHL), age, etiology, duration of CI experience, that are already known to have an influence, were included in the statistical analyses.ResultsThe significant factors were: the pure tone average threshold of the better ear, the brand of device, the percentage of active electrodes, the use of hearing aids (HAs) during the period of pHL, and the duration of mHL.ConclusionsA new model was designed showing a decrease of performance that started during the period of mHL, and became faster during the period of pHL. The use of bilateral HAs slowed down the related central reorganization that is the likely cause of the decreased performance.
Speech is crucial for communication in everyday life. Speech-brain entrainment, the alignment of neural activity to the slow temporal fluctuations (envelope) of acoustic speech input, is a ubiquitous element of current theories of speech processing. Associations between speech-brain entrainment and acoustic speech signal, listening task, and speech intelligibility have been observed repeatedly. However, a methodological bottleneck has prevented so far clarifying whether speech-brain entrainment contributes functionally to (i.e., causes) speech intelligibility or is merely an epiphenomenon of it. To address this long-standing issue, we experimentally manipulated speech-brain entrainment without concomitant acoustic and task-related variations, using a brain stimulation approach that enables modulating listeners' neural activity with transcranial currents carrying speech-envelope information. Results from two experiments involving a cocktail-party-like scenario and a listening situation devoid of aural speech-amplitude envelope input reveal consistent effects on listeners' speech-recognition performance, demonstrating a causal role of speech-brain entrainment in speech intelligibility. Our findings imply that speech-brain entrainment is critical for auditory speech comprehension and suggest that transcranial stimulation with speech-envelope-shaped currents can be utilized to modulate speech comprehension in impaired listening conditions.
When two targets follow each other directly in rapid serial visual presentation (RSVP), they are often identified correctly but reported in the wrong order. These order reversals are commonly explained in terms of the rate at which the two targets are processed, the idea being that the second target can sometimes overtake the first in the race toward conscious awareness. The present study examined whether some of these order reversals might alternatively be due to a mechanism of temporal integration whereby targets appearing closely in time may be merged into a single representation. To test this integration account, we used an attentional blink task in which the two targets could be combined perceptually in a meaningful way such that the conjunction of the two target elements constituted a possible target stimulus itself. The results showed that when targets appeared at Lag 1, observers frequently reported seeing only a single merged target stimulus, and these reports occurred up to approximately three times as often as (real) order reversals. When the possibility to report the integrated percept was removed, order reversals consequently tripled. These results suggest that integration may actually be the primary cause of order reversals in dual-target RSVP tasks.
In normal hearing (NH), the perception of the gender of a speaker is strongly affected by two anatomically related vocal characteristics: the fundamental frequency (F0), related to vocal pitch, and the vocal tract length (VTL), related to the height of the speaker. Previous studies on gender categorization in cochlear implant (CI) users found that performance was variable, with few CI users performing at the level of NH listeners. Data collected with recorded speech produced by multiple talkers suggests that CI users might rely more on F0 and less on VTL than NH listeners. However, because VTL cannot be accurately estimated from recordings, it is difficult to know how VTL contributes to gender categorization. In the present study, speech was synthesized to systematically vary F0, VTL, or both. Gender categorization was measured in CI users, as well as in NH participants listening to unprocessed (only synthesized) and vocoded (and synthesized) speech. Perceptual weights for F0 and VTL were derived from the performance data. With unprocessed speech, NH listeners used both cues (normalized perceptual weight: F0=3.76, VTL=5.56). With vocoded speech, NH listeners still made use of both cues but less efficiently (normalized perceptual weight: F0=1.68, VTL=0.63). CI users relied almost exclusively on F0 while VTL perception was profoundly impaired (normalized perceptual weight: F0=6.88, VTL=0.59). As a result, CI users' gender categorization was abnormal compared to NH listeners. Future CI signal processing should aim to improve the transmission of both F0 cues and VTL cues, as a normal gender categorization may benefit speech understanding in competing talker situations.
The results suggest that listening effort decreases with increased spectral resolution. Moreover, these improvements are best reflected in objective measures of listening effort, such as RTs on a secondary task, rather than intelligibility scores or subjective effort measures.
Objectives:When listening to two competing speakers, normal-hearing (NH) listeners can take advantage of voice differences between the speakers. Users of cochlear implants (CIs) have difficulty in perceiving speech on speech. Previous literature has indicated sensitivity to voice pitch (related to fundamental frequency, F0) to be poor among implant users, while sensitivity to vocal-tract length (VTL; related to the height of the speaker and formant frequencies), the other principal voice characteristic, has not been directly investigated in CIs. A few recent studies evaluated F0 and VTL perception indirectly, through voice gender categorization, which relies on perception of both voice cues. These studies revealed that, contrary to prior literature, CI users seem to rely exclusively on F0 while not utilizing VTL to perform this task. The objective of the present study was to directly and systematically assess raw sensitivity to F0 and VTL differences in CI users to define the extent of the deficit in voice perception.Design:The just-noticeable differences (JNDs) for F0 and VTL were measured in 11 CI listeners using triplets of consonant–vowel syllables in an adaptive three-alternative forced choice method.Results:The results showed that while NH listeners had average JNDs of 1.95 and 1.73 semitones (st) for F0 and VTL, respectively, CI listeners showed JNDs of 9.19 and 7.19 st. These JNDs correspond to differences of 70% in F0 and 52% in VTL. For comparison to the natural range of voices in the population, the F0 JND in CIs remains smaller than the typical male–female F0 difference. However, the average VTL JND in CIs is about twice as large as the typical male–female VTL difference.Conclusions:These findings, thus, directly confirm that CI listeners do not seem to have sufficient access to VTL cues, likely as a result of limited spectral resolution, and, hence, that CI listeners’ voice perception deficit goes beyond poor perception of F0. These results provide a potential common explanation not only for a number of deficits observed in CI listeners, such as voice identification and gender categorization, but also for competing speech perception.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.