We are constantly exposed to our own face and voice, and we identify our own faces and voices as familiar. However, the influence of self-identity upon self-speech perception is still uncertain. Speech perception is a synthesis of both auditory and visual inputs; although we hear our own voice when we speak, we rarely see the dynamic movements of our own face. If visual speech and identity are processed independently, no processing advantage would obtain in viewing one's own highly familiar face. In the present experiment, the relative contributions of facial and vocal inputs to speech perception were evaluated with an audiovisual illusion. Our results indicate that auditory self-speech conveys a processing advantage, whereas visual self-speech does not. The data thereby support a model of visual speech as dynamic movement processed separately from speaker recognition.
WHEN A MUSICAL TONE IS SOUNDED, MOST LISTENERS are unable to identify its pitch by name. Those listeners who can identify pitches are said to have absolute pitch perception (AP). A limited subset of musicians possesses AP, and it has been debated whether musicians' AP interferes with their ability to perceive tonal relationships between pitches, or relative pitch (RP). The present study tested musicians' discrimination of relative pitch categories, or intervals, by placing absolute pitch values in conflict with relative pitch categories. AP listeners perceived intervals categorically, and their judgments were not affected by absolute pitch values. These results indicate that AP listeners do not infer interval identities from the absolute values between tones, and that RP categories are salient musical concepts in both RP and AP musicianship.
The McGurk effect represents a perceptual illusion resulting from the integration of an auditory syllable dubbed onto an incongruous visual syllable. The involuntary and impenetrable nature of the illusion is frequently used to support the multisensory nature of audiovisual speech perception. Here we show that both self-speech and familiarized speech reduce the effect. When self-speech was separated into self-voice and self-face mismatched with different faces and voices, only self-voice weakened the illusion. Thus, a familiar vocal identity automatically confers a processing advantage to multisensory speech, while a familiar facial identity does not. When another group of participants were familiarized with the speakers, participants' ability to take advantage of that familiarization was inversely correlated with their overall susceptibility to the McGurk illusion.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.