2003 IEEE Workshop on Automatic Speech Recognition and Understanding (IEEE Cat. No.03EX721)
DOI: 10.1109/asru.2003.1318446
|View full text |Cite
|
Sign up to set email alerts
|

Recognition of para-linguistic information and its application to spoken dialogue system

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
15
0

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(15 citation statements)
references
References 8 publications
0
15
0
Order By: Relevance
“…Other works deal with the identification of attitudes and intentions of the speaker. For example, Fujie et al (2003) report about the identification of positive/negative attitudes of the speaker, while Maekawa (2000) reports about the classification of paralinguistic items like admiration, suspicion, disappointment and indifference. In Hayashi (1999), paralinguistic items like affirmation, asking again, doubt and hesitation were also considered.…”
Section: Introductionmentioning
confidence: 99%
“…Other works deal with the identification of attitudes and intentions of the speaker. For example, Fujie et al (2003) report about the identification of positive/negative attitudes of the speaker, while Maekawa (2000) reports about the classification of paralinguistic items like admiration, suspicion, disappointment and indifference. In Hayashi (1999), paralinguistic items like affirmation, asking again, doubt and hesitation were also considered.…”
Section: Introductionmentioning
confidence: 99%
“…Nevertheless, a disadvantage of using the eyes as features for motion estimation, is the inability to detect and track faces with sunglasses (commonly worn by visually impaired people). Gunes & Pantic [11] and Fujie et al [12] estimated head motion using optical flow. More precisely, the head region was extracted by skin color segmentation.…”
Section: Related Workmentioning
confidence: 99%
“…A common approach for this is to use HMMs to recognize each gesture. In fact, all the systems described before make use of two or three HMMs with different states [6][7][8][9][11][12][13], e.g., up/down for nodding and left/right for shaking. This approach works well if the system is in a fixed position (i.e., third-party perspective), and is commonly use in applications such as robotics and Human Computer Interaction (HCI), where the system should be able to infer the proper gestures and react accordingly.…”
Section: Related Workmentioning
confidence: 99%
“…Imai et al 2001;Fujie et al 2004), linguistic and paralinguistic cues (e.g. Matsusaka et al 2003;Fujie et al 2005) or emotive vocalization (e.g. Cahn 1990; Abadjieva et al 1993), social touch-based communication (e.g.…”
Section: Introductionmentioning
confidence: 99%