Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue 2015
DOI: 10.18653/v1/w15-4642
|View full text |Cite
|
Sign up to set email alerts
|

Opportunities and Obligations to Take Turns in Collaborative Multi-Party Human-Robot Interaction

Abstract: In this paper we present a data-driven model for detecting opportunities and obligations for a robot to take turns in multi-party discussions about objects. The data used for the model was collected in a public setting, where the robot head Furhat played a collaborative card sorting game together with two users. The model makes a combined detection of addressee and turn-yielding cues, using multi-modal data from voice activity, syntax, prosody, head pose, movement of cards, and dialogue context. The best resul… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
9
1

Relationship

1
9

Authors

Journals

citations
Cited by 22 publications
(22 citation statements)
references
References 29 publications
(38 reference statements)
0
22
0
Order By: Relevance
“…There are some works that investigated other features in speech such as N-gram model [13], dependency structures [14], and the previous turn-taking behaviors [15]. There are also other works that investigated non-verbal features such as respiratory features [16], head pose features [17], and eye-gaze features [18].…”
Section: Turn-taking Predictionmentioning
confidence: 99%
“…There are some works that investigated other features in speech such as N-gram model [13], dependency structures [14], and the previous turn-taking behaviors [15]. There are also other works that investigated non-verbal features such as respiratory features [16], head pose features [17], and eye-gaze features [18].…”
Section: Turn-taking Predictionmentioning
confidence: 99%
“…Given the end of an IPU, the model has to predict whether the speaker is making a pause and "holding" the turn, or whether the speaker is yielding the turn. Various feature sets and machine learning algorithms have been proposed, and tested on both humanhuman and human-machine dialogue data (Meena et al, 2014;Schlangen, 2006;Neiberg and Gustafson, 2011;Johansson and Skantze, 2015;Ferrer et al, 2002;Kawahara et al, 2012).…”
Section: Turn-taking In Spoken Dialoguementioning
confidence: 99%
“…These works and others often focus on a robot's ability to handle a specific aspect of multi-party interactions: receiving and responding to multiple requests [15,26], group detection [28,29], speech recognition [10,11], gesture generation [18], body orientation generation [30], gaze generation [4], etc. Relevant studies in multi-party turn-taking [3,14] use hand-crafted features (e.g., whether someone is speaking, head pose, prosody) to determine when the robot should take a turn, but do not incorporate the contents of speech. The closest multi-party work to ours, [15], uses human-human and human-robot data that was manually labeled to learn low-level submodules for how a bartender robot should interact with multiple customers (e.g., classifying user engagement, or saying pre-defined utterances).…”
Section: Related Workmentioning
confidence: 99%