2019
DOI: 10.7554/elife.48116
|View full text |Cite
|
Sign up to set email alerts
|

The visual speech head start improves perception and reduces superior temporal cortex responses to auditory speech

Abstract: Visual information about speech content from the talker’s mouth is often available before auditory information from the talker's voice. Here we examined perceptual and neural responses to words with and without this visual head start. For both types of words, perception was enhanced by viewing the talker's face, but the enhancement was significantly greater for words with a head start. Neural responses were measured from electrodes implanted over auditory association cortex in the posterior superior temporal g… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

12
46
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
3
2

Relationship

2
6

Authors

Journals

citations
Cited by 41 publications
(59 citation statements)
references
References 62 publications
(77 reference statements)
12
46
0
1
Order By: Relevance
“…We suggest that this benefit is likely to be driven by an increased reliance on the visual articulations which help the listener to understand the noisy speech content by constraining phoneme identity (Karas et al, 2019). In line with this, we also show that the phonemes that most contribute to these results are those that have particularly informative visual articulations (Fig.…”
Section: Enhanced Multisensory Integration Effects At the Phonetic-lesupporting
confidence: 81%
See 1 more Smart Citation
“…We suggest that this benefit is likely to be driven by an increased reliance on the visual articulations which help the listener to understand the noisy speech content by constraining phoneme identity (Karas et al, 2019). In line with this, we also show that the phonemes that most contribute to these results are those that have particularly informative visual articulations (Fig.…”
Section: Enhanced Multisensory Integration Effects At the Phonetic-lesupporting
confidence: 81%
“…One recent perspective (Peelle and Sommers, 2015) suggests that these stages could include an early stage, where visual speech provides temporal cues about the acoustic signal (correlated mode), and a later stage, where visual cues that convey place and manner of articulation could be integrated with acoustic information to constrain lexical selection (complementary mode). Such early-stage integration could be mediated by direct projections from visual cortex that dynamically affect the sensitivity of auditory cortex (Grant and Seitz, 2000;Okada et al, 2013;Tye-Murray et al, 2011;Calvert et al, 1997), whereas for later-stage integration, articulatory visual cues could be combined with acoustic information in supramodal regions such as the STS (Beauchamp et al, 2004;Kayser and Logothetis, 2009;Zhu and Beauchamp, 2017;Karas et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…An important question for further research is determining the mechanisms by which visual speech modulates responses to auditory speech in pSTG/S. One possibility is that the early arrival of visual information influences the responses of populations of neurons in pSTG/S selective for different speech sounds, exciting populations that are compatible with the viewed mouth movements and inhibiting populations that are incompatible with it (6). From a practical perspective, normalizing the response pattern in the service of better comprehension could be accomplished through real-time fMRI neurofeedback (25) or through brain-computer interfaces (26).…”
Section: Discussionmentioning
confidence: 99%
“…In non-human primates, recordings from single neurons in pSTG/S respond to both auditory and visual social communication signals (3)(4)(5). In humans, small populations of neurons in pSTG/S recorded with intracranial electrodes respond to both auditory and visual speech (6,7). While the idea that pSTG/S integrates visual speech information with noisy auditory speech in the service of comprehension seems reasonable, it is supported by only limited empirical evidence.…”
Section: Introductionmentioning
confidence: 99%
“…While this schema was codified in responses to simple auditory beep and visual flash stimuli in anesthetized animals (Stein and Stanford, 2008) it has also been applied to human brain responses to auditory and visual speech recorded with BOLD fMRI, scalp encephalogram (EEG) and intracranial EEG (iEEG) (Besle et al, 2004;van Wassenhove et al, 2005;Besle et al, 2008;Rhone et al, 2016;Karas et al, 2019). In these studies, responses to unisensory and multisensory speech were compared under the assumption that different sensory responses were the main driver of neural activity.…”
Section: Introductionmentioning
confidence: 99%