2015
DOI: 10.1016/j.cortex.2015.04.008
|View full text |Cite
|
Sign up to set email alerts
|

Prediction across sensory modalities: A neurocomputational model of the McGurk effect

Abstract: Audiovisual integrationComputational modeling McGurk effect a b s t r a c tThe McGurk effect is a textbook illustration of the automaticity with which the human brain integrates audio-visual speech. It shows that even incongruent audiovisual (AV) speech stimuli can be combined into percepts that correspond neither to the auditory nor to the visual input, but to a mix of both. Typically, when presented with, e.g., visual /aga/ and acoustic /aba/ we perceive an illusory /ada/. In the inverse situation, however, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
49
0

Year Published

2015
2015
2020
2020

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 39 publications
(54 citation statements)
references
References 52 publications
2
49
0
Order By: Relevance
“…While the mechanisms leading to AV speech fusion are relatively well understood, those leading to AV stimulus combination are still unknown. Based on a previous computational model, we conjectured that AV combination follows from the difficulty to map the auditory and visual physical features in a multisensory space presumably located in the LSTS 28 . AV combination would hence result in a more demanding processing sequence than AV fusion, Figure 1B).…”
Section: Discussionmentioning
confidence: 97%
See 1 more Smart Citation
“…While the mechanisms leading to AV speech fusion are relatively well understood, those leading to AV stimulus combination are still unknown. Based on a previous computational model, we conjectured that AV combination follows from the difficulty to map the auditory and visual physical features in a multisensory space presumably located in the LSTS 28 . AV combination would hence result in a more demanding processing sequence than AV fusion, Figure 1B).…”
Section: Discussionmentioning
confidence: 97%
“…However, according to our predictive model of AV syllable integration 28 an interesting illustration of how predictive coding could apply to AV integration 28,57,58 .…”
Section: The Role Of the Lsts In The Fusion/combination Dynamic Divermentioning
confidence: 95%
“…As described above it proposes that syllables are encoded in terms of the expected amplitudes and variances of audiovisual features. The expected amplitudes were taken from the mean values across 10 productions from a single male speaker (Olasagasti, Bouton, and Giraud 2015), the amplitudes were then normalized by dividing by the highest value for each feature. The remaining parameters in the model, variances and sensory noise levels, were chosen so that the overall categorization results, percentage of /aba/, /ada/ and /aga/ responses to the 6 types of stimuli were qualitatively similar to those reported by Lüttke and colleagues.…”
Section: Model Simulationsmentioning
confidence: 99%
“…The lip and 2 nd formant temporal modulation profiles (M V (t) and M A (t), Fig 1B) were defined as in (Olasagasti, Bouton, and Giraud 2015). Each profile, representing the intervocalic transition between the two vowels in the "a" vocalic context, was modelled with 289 time points.…”
Section: Model Simulationsmentioning
confidence: 99%
See 1 more Smart Citation