2021
DOI: 10.1609/aaai.v35i4.16391
|View full text |Cite
|
Sign up to set email alerts
|

C2F-FWN: Coarse-to-Fine Flow Warping Network for Spatial-Temporal Consistent Motion Transfer

Abstract: Human video motion transfer (HVMT) aims to synthesize videos that one person imitates other persons' actions. Although existing GAN-based HVMT methods have achieved great success, they either fail to preserve appearance details due to the loss of spatial consistency between synthesized and exemplary images, or generate incoherent video results due to the lack of temporal consistency among video frames. In this paper, we propose Coarse-to-Fine Flow Warping Network (C2F-FWN) for spatial-temporal consistent HVMT.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
21
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 12 publications
(21 citation statements)
references
References 25 publications
0
21
0
Order By: Relevance
“…• A Texture Alignment Module is proposed to align the features of the source image and the initially generated image to preserve more details like textures of clothes and edges of the bodies. We conduct extensive experiments on the iPER Dataset [25] and SoloDance Dataset [50]. Experimental results show that our model achieves the state-of-the-art both quantitatively and qualitatively.…”
Section: Introductionmentioning
confidence: 95%
See 4 more Smart Citations
“…• A Texture Alignment Module is proposed to align the features of the source image and the initially generated image to preserve more details like textures of clothes and edges of the bodies. We conduct extensive experiments on the iPER Dataset [25] and SoloDance Dataset [50]. Experimental results show that our model achieves the state-of-the-art both quantitatively and qualitatively.…”
Section: Introductionmentioning
confidence: 95%
“…However, the SMPL model [32] was only suitable for smooth human bodies, it cannot represent the human body with complex clothes. Different from warping the feature of the source image, C2F [50] estimated the optical flow of the clothing regions and directly warped the clothing region according to optical flow. Unfortunately, when the driving pose is greatly different from the source pose, such methods usually failed due to inaccurate flow estimations.…”
Section: Human Video Motion Transfermentioning
confidence: 99%
See 3 more Smart Citations