2007
DOI: 10.1016/j.brainres.2007.03.015
|View full text |Cite
|
Sign up to set email alerts
|

When emotional prosody and semantics dance cheek to cheek: ERP evidence

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

24
128
1

Year Published

2010
2010
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 143 publications
(153 citation statements)
references
References 41 publications
24
128
1
Order By: Relevance
“…Paulmann & Kotz, 2008a;Schirmer & Kotz, 2006), even when this information is not task relevant (Kotz & Paulmann, 2007). These data confirm that prosodic cues ing to both the prime and the target stimuli get automatically activated, since emotionally unrelated targets are processed differently from emotionally related targets in both the 200-and 400-msec prime duration conditions.…”
Section: Discussionsupporting
confidence: 67%
See 1 more Smart Citation
“…Paulmann & Kotz, 2008a;Schirmer & Kotz, 2006), even when this information is not task relevant (Kotz & Paulmann, 2007). These data confirm that prosodic cues ing to both the prime and the target stimuli get automatically activated, since emotionally unrelated targets are processed differently from emotionally related targets in both the 200-and 400-msec prime duration conditions.…”
Section: Discussionsupporting
confidence: 67%
“…According to this perspective, short primes are specifically ignored by participants, and, because they are less activated (i.e., ignored), such primes (and N400-like negativities, such as the N300) reflects the processing of broken emotional context expectations (Bostanov & Kotchoubey, 2004;Kotz & Paulmann, 2007;Paulmann & Kotz, 2008b;Paulmann, Pell, & Kotz, 2008;Schirmer et al, 2002;Zhang et al, 2006). Our data confirm that this emotional context violation response can be elicited by very briefly presented (200-msec) prosodic stimuli, although longer (400-msec) prime durations lead to different effect directions in the N400 (see the Discussion).…”
Section: Reversed Emotional Priming Effectsmentioning
confidence: 99%
“…the P200 component, known to detect emotional salience) is intact (Paulmann et al, 2011). Recent evidence from healthy individuals, where bilateral fronto-striatal activation was found during emotional categorisation tasks on spoken language (especially after elimination of the language's semantic content) (Kotz et al, 2003;Kotz and Paulmann, 2007), further indicates that the basal ganglia play a major function in the processing of affective communication.…”
Section: Basal Gangliamentioning
confidence: 99%
“…Using a crosssplicing technique, a leading phrase of a sentence was crossspliced with the main stem of a sentence either congruent or incongruent in prosody with the leading phrase. The onset of the crosssplicing point of the vocal expression in the main sentence elicited a larger negativity (350-550 ms) for a mismatch in both voice and lexico-semantics and a larger more right-hemispheric distributed positivity (600-950 ms) for a mismatch in voice only (pseudouterances: [3]; uterances with no emotional lexical items: [1]). The negativity suggested an efort of integrating the emotional information in both vocal and semantic channel with the context.…”
Section: Neurophysiological Studies On Basic Vocal Emotion In Speechmentioning
confidence: 97%
“…), which roughly correspond to each hypothesized processing stage [1,2]. Studies with event-related potentials (ERPs), which focused on the analysis of averaged electrophysiological response to a certain vocal or speech event, have enlightened neurocognitive processes at a ine-grained temporal scale.…”
Section: Introductionmentioning
confidence: 99%