2005
DOI: 10.1007/s11265-005-4150-4
|View full text |Cite
|
Sign up to set email alerts
|

A Speech-Centric Perspective for Human-Computer Interface: A Case Study

Abstract: Speech technology has been playing a central role in enhancing human-machine interactions, especially for small devices for which graphical user interface has obvious limitations. The speech-centric perspective for human-computer interface advanced in this paper derives from the view that speech is the only natural and expressive modality to enable people to access information from and to interact with any device. In this paper, we describe some recent work conducted at Microsoft Research, aimed at the develop… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2008
2008
2008
2008

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(5 citation statements)
references
References 16 publications
(11 reference statements)
0
5
0
Order By: Relevance
“…This feature makes the late fusion easier to scale up to more modalities in the future than the early fusion. The architecture shown in Figure 1 utilizes the late fusion approach that has been widely adopted, for example, by a variety of systems including Put-That-There (Bolt, 1980), MapPointS (Deng & Yu, 2005), MiPad (Huang et al, 2001), ShopTalk (Cohen, et al, 1989), QuickSet (Cohen, Johnston, McGee, Oviatt, Pittman, Smith, et al, 1997), CUBRICON (Neal & Shapiro, 1991), Virtual World (Codella, Jalili, Koved, Lewis, Ling, Lipscomb, et al, 1992), Finger-Pointer (Fukumoto et al, 1994), VisualMan (Wang, 1995), and Jeanie (Vo & Wood, 1996).…”
Section: A Generic Mui Architecturementioning
confidence: 99%
See 4 more Smart Citations
“…This feature makes the late fusion easier to scale up to more modalities in the future than the early fusion. The architecture shown in Figure 1 utilizes the late fusion approach that has been widely adopted, for example, by a variety of systems including Put-That-There (Bolt, 1980), MapPointS (Deng & Yu, 2005), MiPad (Huang et al, 2001), ShopTalk (Cohen, et al, 1989), QuickSet (Cohen, Johnston, McGee, Oviatt, Pittman, Smith, et al, 1997), CUBRICON (Neal & Shapiro, 1991), Virtual World (Codella, Jalili, Koved, Lewis, Ling, Lipscomb, et al, 1992), Finger-Pointer (Fukumoto et al, 1994), VisualMan (Wang, 1995), and Jeanie (Vo & Wood, 1996).…”
Section: A Generic Mui Architecturementioning
confidence: 99%
“…An even more effective language model would weight different names differently, depending on the frequencies the user exchanged e-mail with the person, and the recentness of the interaction (Yu, Wang, Mahajan, Mau, & Acero, 2003). Another example of constructing the language model based on the context and user information is described in the speech enabled MapPoint (Deng & Yu, 2005). Without context information, the speech recognizer needs to load all location names and business names in the North America.…”
Section: Context-aware Language Modelmentioning
confidence: 99%
See 3 more Smart Citations