SEMDIAL 2017 (SaarDial) Workshop on the Semantics and Pragmatics of Dialogue 2017
DOI: 10.21437/semdial.2017-5
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Coreference Resolution for Exploratory Data Visualization Dialogue: Context-Based Annotation and Gesture Identification

Abstract: The goals of our work are twofold: gain insight into how humans interact with complex data and visualizations thereof in order to make discoveries; and use our findings to develop a dialogue system for exploring data visualizations. Crucial to both goals is understanding and modeling of multimodal referential expressions, in particular those that include deictic gestures. In this paper, we discuss how context information affects the interpretation of requests and their attendant referring expressions in our da… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 17 publications
0
5
0
Order By: Relevance
“…Some recent work has started investigating the potential of building dialogue systems that can help users efficiently explore data through visualizations (Kumar et al, 2017).…”
Section: Related Workmentioning
confidence: 99%
“…Some recent work has started investigating the potential of building dialogue systems that can help users efficiently explore data through visualizations (Kumar et al, 2017).…”
Section: Related Workmentioning
confidence: 99%
“…Our work complements this analysis by observing similar tasks that spanned many views and utilized large display areas. Our work expands upon Aurisano et al [AKG*b, AKG*a], which presents preliminary analysis of this data, and Kumar et al [KADE*16, KDEA*], which examines utterances and gestures from a natural language processing perspective.…”
Section: Related Workmentioning
confidence: 88%
“…To make multimodal interaction more smooth, the system should have a declarative representation for all these potential referents and find the best match. Articulate2 [120] addresses this issue by leveraging Kinect to detect deictic gestures on the virtual touch screen in front of a large display. If referring expressions are detected and a gesture has been detected by Kinect, information about any objects pointed to by the user will be stored and then the system can find the best match between properties of each relevant entity.…”
Section: Co-reference Resolutionmentioning
confidence: 99%