2021
DOI: 10.48550/arxiv.2108.04325
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

AnyoneNet: Synchronized Speech and Talking Head Generation for Arbitrary Person

Xinsheng Wang,
Qicong Xie,
Jihua Zhu
et al.

Abstract: Automatically generating videos in which synthesized speech is synchronized with lip movements in a talking head has great potential in many human-computer interaction scenarios. In this paper, we present an automatic method to generate synchronized speech and talking-head videos on the basis of text and a single face image of an arbitrary person as input. In contrast to previous text-driven talking head generation methods, which can only synthesize the voice of a specific person, the proposed method is capabl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 55 publications
0
1
0
Order By: Relevance
“…Traditional talking head generation models [49,9,47,42,60,12] focus on synthesizing audio-synchronized lip motion, but only generate lip motion with fixed head poses. To address this issue, some recent works consider personalized attributes [52,55,58,50,48,7]. However, these methods [52,55,58] generate personalized information with a deterministic model and the results are short of diversity, leading to a repetitive pattern.…”
Section: Introductionmentioning
confidence: 99%
“…Traditional talking head generation models [49,9,47,42,60,12] focus on synthesizing audio-synchronized lip motion, but only generate lip motion with fixed head poses. To address this issue, some recent works consider personalized attributes [52,55,58,50,48,7]. However, these methods [52,55,58] generate personalized information with a deterministic model and the results are short of diversity, leading to a repetitive pattern.…”
Section: Introductionmentioning
confidence: 99%