2003
DOI: 10.1111/1467-8659.t01-2-00711
|View full text |Cite
|
Sign up to set email alerts
|

Visyllable Based Speech Animation

Abstract: Visemes are visual counterpart of phonemes. Traditionally, the speech animation of 3D synthetic faces involvesextraction of visemes from input speech followed by the application of co‐articulation rules to generate realisticanimation. In this paper, we take a novel approach for speech animation — using visyllables, the visual counterpartof syllables. The approach results into a concatenative visyllable based speech animation system. The key contributionof this paper lies in two main areas. Firstly, we define a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
33
0
1

Year Published

2005
2005
2014
2014

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 52 publications
(34 citation statements)
references
References 11 publications
(14 reference statements)
0
33
0
1
Order By: Relevance
“…The temporal aspects of facial performances are very important for synthesis of new facial animations from speech [Bregler et al 1997;Brand 1999;Ezzat et al 2002;Kshirsagar and Thalmann 2003;Cao et al 2004;Ma et al 2004;Deng et al 2005]. Most of these approaches record facial motion of speaking subjects and then recombine the recorded facial motion from learned parametric models to synthesize new facial motion.…”
Section: Related Workmentioning
confidence: 99%
“…The temporal aspects of facial performances are very important for synthesis of new facial animations from speech [Bregler et al 1997;Brand 1999;Ezzat et al 2002;Kshirsagar and Thalmann 2003;Cao et al 2004;Ma et al 2004;Deng et al 2005]. Most of these approaches record facial motion of speaking subjects and then recombine the recorded facial motion from learned parametric models to synthesize new facial motion.…”
Section: Related Workmentioning
confidence: 99%
“…The motion synthesis is mainly on the concatenation of visual units. [2][3][4]6,7 Ezzat et al 2 employ a variant of Multidimensional Morphable Model (MMM) to embed mouth configurations. HMM is employed to model the probabilistic state machine for speech animation.…”
Section: Related Workmentioning
confidence: 99%
“…Currently, the 3D facial motions can be captured real-time with opticalbased devices. Given the large volume of dataset, most 3D speech motion synthesis algorithms are based on the concatenation of captured motion segments and the variable transition models [1][2][3][4][5] to handle the coarticulation. Hitherto, little efforts addressed the inside dynamic nature of the pronunciation unit, e.g., what dynamics inside the visyllable is and how the facial shape varies throughout one visyllable.…”
Section: Introductionmentioning
confidence: 99%
“…[Byun and Badler 2002] modified the MPEG-4 Facial Animation Parameters (FAPs) [Moving Picture Experts Group 1998] to add expressiveness, [Kshirsagar and Magnenat-Thalmann 2003] used PCA to deform the mouth during speech, and [Chai et al 2003] used facial tracking to drive animations from a motion capture database. [Wang et al 2004] used a multiresolution deformable mesh to track facial motion, and a low dimensional embedding technique to learn expression style.…”
Section: Related Workmentioning
confidence: 99%