2002
DOI: 10.1109/tsa.2002.804538
|View full text |Cite
|
Sign up to set email alerts
|

Distributed speech processing in miPad's multimodal user interface

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
28
0

Year Published

2004
2004
2018
2018

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 42 publications
(28 citation statements)
references
References 5 publications
0
28
0
Order By: Relevance
“…Before deep learning methods were adopted, there had already been numerous efforts in multimodal and multitask learning. For example, a prototype called MiPad for multimodal interactions involving capturing, leaning, coordinating, and rendering a mix of speech, touch, and visual information was developed and reported in [113,164]. In [165,166], mixed sources of information from multiple-sensory microphones with separate bone-conductive and air-born paths were exploited to de-noise speech.…”
Section: B) a Selected Review On Deep Learning For Multimodal Processingmentioning
confidence: 99%
“…Before deep learning methods were adopted, there had already been numerous efforts in multimodal and multitask learning. For example, a prototype called MiPad for multimodal interactions involving capturing, leaning, coordinating, and rendering a mix of speech, touch, and visual information was developed and reported in [113,164]. In [165,166], mixed sources of information from multiple-sensory microphones with separate bone-conductive and air-born paths were exploited to de-noise speech.…”
Section: B) a Selected Review On Deep Learning For Multimodal Processingmentioning
confidence: 99%
“…MiPad is a prototype of wireless Personal Digital Assistant (PDA) that enables users to accomplish many common tasks using a multimodal spoken language interface (speech + pen + display). MiPad, as a case study for speech-centric multimodal HCI application, has been described in detail in our recent publication [2]. In this paper, we will present a second case study based on a new system built within our research group more recently, called MapPointS.…”
Section: Introductionmentioning
confidence: 99%
“…Many prototype systems have also been built based on the use of multiple modalities [1,2,7,9,14], most of which have focused on the special advantage of the speech input for mobile or wireless computing as in multimodal PDA's. Both of our prototype systems, MiPad and MapPointS, have such mobile computing in the special design consideration.…”
Section: Introductionmentioning
confidence: 99%
“…The total cost for a cluster system including maintenance is obviously lower than that of a DSP-based system in research and development stage. From the point of view of acoustical application using network communication, distributed speech processing for a personal digital assistant (PDA) is discussed [6]. The literature proposes that the speech signal is transmitted from the PDA to the remote server for automatic speech recognition.…”
Section: Introductionmentioning
confidence: 99%