Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
2017 IEEE International Conference on Multimedia and Expo (ICME) 2017
DOI: 10.1109/icme.2017.8019546
|View full text |Cite
|
Sign up to set email alerts
|

Visual speech synthesis from 3D mesh sequences driven by combined speech features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…They focused on the tip of the tongue between the teeth or the back of the tongue, but this type of approach requires sufficient data collected from the subject in advance to produce animation results. Kuhnke and Ostermann collected a sequence of 3D mesh data along the phoneme label by capturing 3D facial movement and recording voice data at the same time [11]. With a regression-based method, they presented a combination method of several speech features for better performance.…”
Section: Text-driven Speech Animation Generationmentioning
confidence: 99%
See 1 more Smart Citation
“…They focused on the tip of the tongue between the teeth or the back of the tongue, but this type of approach requires sufficient data collected from the subject in advance to produce animation results. Kuhnke and Ostermann collected a sequence of 3D mesh data along the phoneme label by capturing 3D facial movement and recording voice data at the same time [11]. With a regression-based method, they presented a combination method of several speech features for better performance.…”
Section: Text-driven Speech Animation Generationmentioning
confidence: 99%
“…It should be able to express the shape of the lips that is precisely synchronized with the speaking voice. Numerous studies have presented ways to create visual speech animation with the speech track [1,2,3,4,5,6,7,8,9,10,11] while other approaches have focused on simulating facial movements from a set of physical properties [12,13,14,15] or synthesizing emotional expressions from given facial models [16,17,18,19,20]. For more natural and realistic facial animation, an explicit solution is needed to combine the lip movements and facial expressions into one animation sequence.…”
Section: Introductionmentioning
confidence: 99%
“…ECAs are those CAs that can facilitate full virtual body and the available embodiment in order to incorporate humanlike responses. The ECA technology ranges from chatbots and 2D/3D realizations in a form of talking heads [22][23][24] to fully articulated embodied conversational agents engaged in various concepts of HMI, including sign language [25], storytelling [26], companions [27], and virtual hosts within user interfaces, and even used as moderators of various concepts in ambient-assisted living environments [28][29][30][31][32].…”
Section: Introductionmentioning
confidence: 99%