2011
DOI: 10.1016/j.specom.2010.11.006
|View full text |Cite
|
Sign up to set email alerts
|

Achieving rapport with turn-by-turn, user-responsive emotional coloring

Abstract: People in dialog use a rich set of nonverbal behaviors, including variations in the prosody of their utterances. Such behaviors, often emotion-related, call for appropriate responses, but today's spoken dialog systems lack the ability to do this. Recent work has shown how to recognize user emotions from prosody and how to express system-side emotions with prosody, but demonstrations of how to combine these functions to improve the user experience have been lacking. Working with a corpus of conversations with s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
29
0

Year Published

2012
2012
2018
2018

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 80 publications
(31 citation statements)
references
References 41 publications
2
29
0
Order By: Relevance
“…A similar finding was found in [38]. The NEMOHIFI version's response (NRG content) was considerably lengthier, and the addition of prosody has made it even more lengthy than in previous experiments.…”
Section: Resultssupporting
confidence: 86%
“…A similar finding was found in [38]. The NEMOHIFI version's response (NRG content) was considerably lengthier, and the addition of prosody has made it even more lengthy than in previous experiments.…”
Section: Resultssupporting
confidence: 86%
“…They quantitatively measure the rhythm entrainment between speakers as the latency of the first pitch accent after a turn exchange divided by the rate of pitch accents in the utterance preceding the turn exchange. Finally, [78] proposed using a machine learning algorithm to predict the emotional coloring (valence, activation, power) of an utterance based on the emotional coloring of the previous utterance.…”
Section: Fully Automatic Measures Of Move-ment Synchronymentioning
confidence: 99%
“…For example, in (Bevacqua et al, 2010) vocal cues are used to generate backchannels (i.e., non-intrusive signals provided during the speaker's turn). Acosta and Ward proposed a system that uses speech and prosody variation to build rapport between human and agent (Acosta and Ward, 2011), and Cavazza et al used vocal signals for character-based interactive storytelling (Cavazza et al, 2009). Furthermore, the virtual human SimSensei Kiosk uses voice, speech and other features to analyse user emotions in the context of healthcare decision support (DeVault et al, 2014).…”
Section: Emotions In Vocal Signalsmentioning
confidence: 99%