2013
DOI: 10.1016/j.neuroimage.2013.05.002
|View full text |Cite
|
Sign up to set email alerts
|

Mean-based neural coding of voices

Abstract: The social significance of recognizing the person who talks to us is obvious, but the neural mechanisms that mediate talker identification are unclear. Regions along the bilateral superior temporal sulcus (STS) and the inferior frontal cortex (IFC) of the human brain are selective for voices, and they are sensitive to rapid voice changes. Although it has been proposed that voice recognition is supported by prototype-centered voice representations, the involvement of these category-selective cortical regions in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
32
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 34 publications
(34 citation statements)
references
References 54 publications
(91 reference statements)
2
32
0
Order By: Relevance
“…Recall that the typicality analysis collapsed both Long and Short VOT variants for both talkers (each of which might be perceived as ‘typical’ or ‘atypical’ of a given talker according to the counterbalanced group assignment), so differences in activation seen within this contrast cannot be attributed to surface-level properties of the stimuli. Notably, right posterior temporal and parietal regions have been linked in lesion studies (Van Lancker et al, 1988, 1989) and imaging studies (Andics et al, 2010, 2013) to access to talker identity. This region also abuts a slightly more ventral MTG area that was found to be responsive in a previous study to talker-specific phonetic variability when that variability takes the form of an ambiguous phoneme inserted in a biasing lexical context (Myers & Mesite, 2014).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Recall that the typicality analysis collapsed both Long and Short VOT variants for both talkers (each of which might be perceived as ‘typical’ or ‘atypical’ of a given talker according to the counterbalanced group assignment), so differences in activation seen within this contrast cannot be attributed to surface-level properties of the stimuli. Notably, right posterior temporal and parietal regions have been linked in lesion studies (Van Lancker et al, 1988, 1989) and imaging studies (Andics et al, 2010, 2013) to access to talker identity. This region also abuts a slightly more ventral MTG area that was found to be responsive in a previous study to talker-specific phonetic variability when that variability takes the form of an ambiguous phoneme inserted in a biasing lexical context (Myers & Mesite, 2014).…”
Section: Discussionmentioning
confidence: 99%
“…Imaging studies have further corroborated the separation between regions that are sensitive to the acoustics of the voice—and thus could be used for discriminating between talkers—and those responsible for mapping voice acoustics to an individual identity which can be used for talker identification (von Kriegstein, et al, 2003). In particular, while voice acoustics may be processed in bilateral temporal regions (specifically the superior temporal sulcus or STS), imaging studies have sited vocal identity processing (or access to familiar voices) in the anterior right temporal lobe rather than the right posterior region implicated in lesion studies (Andics et al, 2010; Andics, McQueen, & Petersson, 2013; Belin & Zatorre, 2003; Campanella & Belin, 2007). …”
Section: Introductionmentioning
confidence: 99%
“…Thus normal processing of voices involves increasingly more abstract representations of speaker identity, independent of other (e.g., acoustic) features (Warren et al 2006). Current evidence also suggests that individual voices are coded relative to how different they are from a prototypical, or average, voice (i.e., a ‘prior’; Andics et al 2013; Andics et al 2010; Latinus et al 2013), which indicates that both prediction and prediction error signals are routinely implemented in the perception of speaker identity.…”
Section: An Auditory Processing Stream Frameworkmentioning
confidence: 99%
“…Based on the previous literature, we expected in the crossmodal condition a decrease in activation in samespeaker (person-congruent) compared with differentspeaker (person-incongruent) trials in face-sensitive areas of the fusiform gyrus [Grill-Spector et al, 2004;Shah et al, 2001] and in voice-sensitive areas along the STS [Andics et al, 2013b;Joassin et al, 2011;Latinus et al, 2011] and the inferior frontal gyrus (IFG, [Andics et al, 2013a,b;Latinus et al, 2011]). The S2 stimulus (the target) was a human voice in all conditions.…”
Section: Introductionmentioning
confidence: 99%
“…We further manipulated the congruency between the prime and the target, that is, whether the prime and the target belonged to the same speaker (person-congruent) or to different speakers (person-incongruent). Based on the previous literature, we expected in the crossmodal condition a decrease in activation in samespeaker (person-congruent) compared with differentspeaker (person-incongruent) trials in face-sensitive areas of the fusiform gyrus [Grill-Spector et al, 2004;Shah et al, 2001] and in voice-sensitive areas along the STS [Andics et al, 2013b;Joassin et al, 2011;Latinus et al, 2011] and the inferior frontal gyrus (IFG, [Andics et al, 2013a,b;Latinus et al, 2011]). We further predicted a similar decline of the BOLD signal in supramodal brain regions which have been previously reported to be activated during the integration of human faces and voices, specifically the pSTS [Blank et al, 2011;Joassin et al, 2011;Klasen et al, 2011;Watson et al, 2013Watson et al, , 2014a and the angular gyrus [Joassin et al, 2011;Klasen et al, 2011;M€ uller et al, 2011].…”
Section: Introductionmentioning
confidence: 99%