2005
DOI: 10.1007/11508373_7
|View full text |Cite
|
Sign up to set email alerts
|

Utilizing Visual Attention for Cross-Modal Coreference Interpretation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2006
2006
2013
2013

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(10 citation statements)
references
References 19 publications
0
9
0
Order By: Relevance
“…Emerging technologies ranging from conversational agents that interact directly with humans on collaborative physical tasks [11], to VR and AR systems that attempt to comprehend spoken references to objects in the environment [8,28], to videomediated communication systems that track conversation and automatically adapt their views based on what the pairs need to see [26,29], would benefit from more advanced computational models of human referring behavior and an understanding of the ways in which context influences collaborative reference. Furthermore, new technologies that provide lightweight mobile eye tracking capabilities [7] are quickly becoming available as platforms for collaborative technologies that reside in everyday physical settings away from the desktop.…”
Section: Reference and Technologymentioning
confidence: 99%
“…Emerging technologies ranging from conversational agents that interact directly with humans on collaborative physical tasks [11], to VR and AR systems that attempt to comprehend spoken references to objects in the environment [8,28], to videomediated communication systems that track conversation and automatically adapt their views based on what the pairs need to see [26,29], would benefit from more advanced computational models of human referring behavior and an understanding of the ways in which context influences collaborative reference. Furthermore, new technologies that provide lightweight mobile eye tracking capabilities [7] are quickly becoming available as platforms for collaborative technologies that reside in everyday physical settings away from the desktop.…”
Section: Reference and Technologymentioning
confidence: 99%
“…Eye gaze has been explored in automated language understanding such as speech recognition [4,14], reference resolution [3,13], and recently for word acquisition [10,22]. Given speech paired with eye gaze information and video images, a translation model was used to acquire words by associating acoustic phone sequences with visual representations of objects and actions [22].…”
Section: Related Workmentioning
confidence: 99%
“…Eye gaze as a modality in multimodal interaction goes beyond the function of pointing. In different speech and eye gaze systems, eye gaze has been explored for the purpose of mutual disambiguation (Tanaka, 1999;Zhang, Imamiya, Go, & Mao, 2004), as a complement to the speech channel for reference resolution (Campana, Baldridge, Dowding, Hockey, Remington, & Stone, 2001;Kaur, Termaine, Huang, Wilder, Gacovski, Flippo, & Mantravadi, 2003;Prasov & Chai, 2008;Byron, Mampilly, Sharma, & Xu, 2005) and speech recognition (Cooke, 2006;Qu & Chai, 2007), and for managing human-computer dialogue (Qvarfordt & Zhai, 2005).…”
Section: Related Workmentioning
confidence: 99%