Interspeech 2008 2008
DOI: 10.21437/interspeech.2008-591
|View full text |Cite
|
Sign up to set email alerts
|

Speech-driven lip motion generation with a trajectory HMM

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
3
0

Year Published

2012
2012
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 29 publications
(5 citation statements)
references
References 9 publications
0
3
0
Order By: Relevance
“…It is important to note that efforts to create systems that transform acoustic speech to animated faces predate DNNs’ widespread use. For example, statistical machine learning (ML) methods such as Hidden Markov models (HMM) have been used previously to generate moving mouth animations from either text or speech audio (Al Moubayed et al, 2008; Hofer et al, 2008; Masuko et al, 1998; Schabus et al, 2013; Tamura et al, 1998). Other ML methods such as QR factorization (Lucero et al, 2006) and artificial neural networks (ANN) (Massaro et al, 1999) have also been used for similar purposes.…”
Section: Synthetic Talking Face Generation Using Deep Neural Network ...mentioning
confidence: 99%
“…It is important to note that efforts to create systems that transform acoustic speech to animated faces predate DNNs’ widespread use. For example, statistical machine learning (ML) methods such as Hidden Markov models (HMM) have been used previously to generate moving mouth animations from either text or speech audio (Al Moubayed et al, 2008; Hofer et al, 2008; Masuko et al, 1998; Schabus et al, 2013; Tamura et al, 1998). Other ML methods such as QR factorization (Lucero et al, 2006) and artificial neural networks (ANN) (Massaro et al, 1999) have also been used for similar purposes.…”
Section: Synthetic Talking Face Generation Using Deep Neural Network ...mentioning
confidence: 99%
“…A rich set of other methods [29] were designed so as to produce expression-guided speech videos. In [30], the authors proposed to intelligently predict lip-based movement trajectory using human speech. The designed system accurately calculates human lip movements from the original human…”
Section: B Facial Animation Video Driven By Speechmentioning
confidence: 99%
“…In this subsection, our designed animation system is compared with the facial systems proposed by Deng et al [28], Kshirsagar et al [29], Hofer et al [30], and Zoric et al [40] respectively. Noticeably, either accuracy or ranking is an optimal choice for this task.…”
Section: ) Morphing-based Facial Animation Videomentioning
confidence: 99%
“…A rich set of other methods [27] were designed so as to produce expression-guided speech videos. In [28], the authors proposed to intelligently predict lip-based movement trajectory using human speech. The designed system accurately calculates human lip movements from the original human speech.…”
Section: B Facial Animation Video Driven By Speechmentioning
confidence: 99%
“…In this subsection, our designed animation system is compared with the facial systems proposed by Deng et al [26], Kshirsagar et al [27], Hofer et al [28], and Zoric et al [29] respectively. Noticeably, either accuracy or ranking is an optimal choice for this task.…”
Section: ) Morphing-based Facial Animation Videomentioning
confidence: 99%