2010 7th International Symposium on Chinese Spoken Language Processing 2010
DOI: 10.1109/iscslp.2010.5684832
|View full text |Cite
|
Sign up to set email alerts
|

Development of an articulatory visual-speech synthesizer to support language learning

Abstract: -This paper presents a two-dimensional (2D) visualspeech synthesizer to support language learning. A visual-speech synthesizer animates the human articulators in synchronization with speech signals, e.g., output from a text-to-speech synthesizer. A visual-speech animation can offer a concrete illustration to the language learners on how to move and where to place the articulators when pronouncing a phoneme. We adopt a 2D vector-based viseme models and compiled a collection of visemes to cover the articulation … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2011
2011
2019
2019

Publication Types

Select...
4
2

Relationship

3
3

Authors

Journals

citations
Cited by 8 publications
(12 citation statements)
references
References 1 publication
0
12
0
Order By: Relevance
“…Our exaggerated feedback is implemented on the previous visual-speech synthesizer reported in [9][10], which visualized pronunciation movement from midsagittal and front views of the vocal tract. It focuses on providing corrective feedback and can offer a reliable visualization for coarticulation.…”
Section: The Exaggeration Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Our exaggerated feedback is implemented on the previous visual-speech synthesizer reported in [9][10], which visualized pronunciation movement from midsagittal and front views of the vocal tract. It focuses on providing corrective feedback and can offer a reliable visualization for coarticulation.…”
Section: The Exaggeration Methodsmentioning
confidence: 99%
“…In our visual-speech synthesis system, each phoneme is assigned to two visemes as the key frames for animation generation [9][10]. Each viseme can be assumed as the representation of a key articulatory action.…”
Section: Realization Of Visual Exaggerationmentioning
confidence: 99%
See 1 more Smart Citation
“…Learning approaches to pronunciation can be roughly divided into two types: phonics training and whole-word training. 13 The phonics training approach emphasizes phoneme training and aims to correct phoneme-level errors. 14 In contrast, whole-word training emphasizes meaning to encourage memorization by students.…”
Section: Introductionmentioning
confidence: 99%
“…WASAY generates synchronized animations of the speech articulators in the midsagittal and the front views. The initial implementation of WASAY [6] uses context-independent visemes, and blending such visemes does not offer a reliable visualization for coarticulation.…”
Section: Introductionmentioning
confidence: 99%