formance in the AV condition for all 3 stimulus types surpassed what would be predicted by simply adding the scores obtained in the A and V conditions. On average, for example, the subjects recognized about 10% of the words in the Iowa Sentence Test 4 in a V condition, about 40% of the words in an A condition with a background of 6-talker babble, and nearly 80% of the words in an AV condition, also with a background of 6-talker babble. A similar superadditive effect was observed for the group of 44 older adults (ages 65 years and older) who were also included in the study and for the consonant and word tests.One reason for this superadditive effect is the complementary nature of the auditory and visual speech signals.1,5 For example, cues about nasality and voicing are typically conveyed very well by the auditory signal, even in adverse listening situations, whereas the visual signal does not convey them at all, even in the best of viewing conditions. On the other hand, cues about place of articulation are conveyed by the visual signal but not very well by a degraded M ost people, whether they have normal or impaired hearing, can recognize speech better when they can both see and hear the talker than when listening alone.1,2 Often, the advantage of supplementing listening with watching is more than additive. For instance, Sommers et al 3 tested 38 young adults (between the ages of 18 and 25 years) in an audition-only (A), vision-only (V), and audition-plus-vision (AV) condition with 3 stimuli types-consonants, words, and sentences. The perMuch evidence suggests that the mental lexicon is organized into auditory neighborhoods, with words that are phonologically similar belonging to the same neighborhood. In this investigation, we considered the existence of visual neighborhoods. When a receiver watches someone speak a word, a neighborhood of homophenes (ie, words that look alike on the face, such as pat and bat) is activated. The simultaneous activation of a word's auditory and visual neighborhoods may, in part, account for why individuals recognize speech better in an auditory-visual condition than what would be predicted by their performance in audition-only and vision-only conditions. A word test was administered to 3 groups of participants in audition-only, vision-only, and auditory-visual conditions, in the presence of 6-talker babble. Test words with sparse visual neighborhoods were recognized more accurately than words with dense neighborhoods in a vision-only condition. Densities of both the acoustic and visual neighborhoods as well as their intersection overlap were predictive of how well the test words were recognized in the auditory-visual condition. These results suggest that visual neighborhoods exist and that they affect auditory-visual speech perception. One implication is that in the presence of dual sensory impairment, the boundaries of both acoustic and visual neighborhoods may shift, adversely affecting speech recognition.