Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-2067
|View full text |Cite
|
Sign up to set email alerts
|

Engagement Recognition in Spoken Dialogue via Neural Network by Aggregating Different Annotators' Models

Abstract: This paper addresses engagement recognition based on four multimodal listener behaviors-backchannels, laughing, eyegaze, and head nodding. Engagement is an indicator of how much a user is interested in the current dialogue. Multiple third-party annotators give ground truth labels of engagement in a human-robot interaction corpus. Since perception of engagement is subjective, the annotations are sometimes different between individual annotators. Conventional methods directly use integrated labels, such as those… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(9 citation statements)
references
References 36 publications
0
8
0
Order By: Relevance
“…Engagement Metrics Engagement is a substantial metric that shows user willingness to continue conversing with the system (Yu, Aoki, and Woodruff 2004;Ma 2018;Inoue et al 2018) and has been studied in the context of dialogue systems (Yu et al 2016;Zhang et al 2018;See et al 2019). Many researchers have considered engagement as a useful metric toward achieving better dialogue systems (Yu et al 2016;Zhang et al 2018).…”
Section: Relevance Metricsmentioning
confidence: 99%
“…Engagement Metrics Engagement is a substantial metric that shows user willingness to continue conversing with the system (Yu, Aoki, and Woodruff 2004;Ma 2018;Inoue et al 2018) and has been studied in the context of dialogue systems (Yu et al 2016;Zhang et al 2018;See et al 2019). Many researchers have considered engagement as a useful metric toward achieving better dialogue systems (Yu et al 2016;Zhang et al 2018).…”
Section: Relevance Metricsmentioning
confidence: 99%
“…Salam et al [43] used person to detect both individual and group engagement. [26,27] combined different aspects like backchannels, eye gaze, head nodding-based features to detect engagement level. Ben et al [4] combined several attributes like speech and facial expressions, gaze and head motion, distance to robot to identify disengagement.…”
Section: Related Workmentioning
confidence: 99%
“…While most of the work has been done to identify basic prototypic emotions, there have been some work in building automated systems to detect more complex emotional and cognitive states. For example, researchers in the education domain have attempted to build automatic models to detect student engagement, a phenomenon inversely related to boredom, from face [6,45] and speech [21]. Researchers have also attempted to model mind-wandering, another phenomenon related to boredom, using eye-gaze behavior [20,29] and physiological signals [5].…”
Section: Boredom: Elicitation and Mitigationmentioning
confidence: 99%