2006 IEEE/ACM International Symposium on Mixed and Augmented Reality 2006
DOI: 10.1109/ismar.2006.297812
|View full text |Cite
|
Sign up to set email alerts
|

"Move the couch where?" : developing an augmented reality multimodal interface

Abstract: This paper describes an augmented reality (AR) multimodal interface that uses speech and paddle gestures for interaction. The application allows users to intuitively arrange virtual furniture in a virtual room using a combination of speech and gestures from a real paddle. Unlike other multimodal AR applications, the multimodal fusion is based on the combination of time-based and semantic techniques to disambiguate a users speech and gesture input. We describe our AR multimodal interface architecture and discus… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0
1

Year Published

2008
2008
2024
2024

Publication Types

Select...
4
3
3

Relationship

3
7

Authors

Journals

citations
Cited by 40 publications
(18 citation statements)
references
References 18 publications
0
17
0
1
Order By: Relevance
“…The user had to wear a data glove and the researchers did not conduct user studies to explore the effectiveness of SenseShapes. Irawati et al [15] has developed a computer vision based multimodal AR system by adding speech input to the VOMAR furniture arranging application [16]. The final system allowed a user to pick and place virtual furniture in an AR scene using a combination of paddle gestures and speech commands.…”
Section: Related Workmentioning
confidence: 99%
“…The user had to wear a data glove and the researchers did not conduct user studies to explore the effectiveness of SenseShapes. Irawati et al [15] has developed a computer vision based multimodal AR system by adding speech input to the VOMAR furniture arranging application [16]. The final system allowed a user to pick and place virtual furniture in an AR scene using a combination of paddle gestures and speech commands.…”
Section: Related Workmentioning
confidence: 99%
“…This report is basically based on the application that is to be introduced which will be a mobile application that intends to provide wealthier experiences by overlying words or virtual objects over the scene observed through a camera. The mobile application is a real time visual translator for Bahasa language to English that will be embedded on mobile phones [10]- [12].…”
Section: A Mobile Learningmentioning
confidence: 99%
“…Their method was robust against partial occlusions of the patterns. Irawati et al [39] showed an extension of Kato's VOMAR interface that fused speech with paddle-based gesture input using time-based and semantic techniques. The overall goal of these new interaction techniques is to enable manipulating AR content to be as easy as interacting with real objects.…”
Section: 21mentioning
confidence: 99%