Proceedings of the 28th ACM International Conference on Multimedia 2020
DOI: 10.1145/3394171.3413932
|View full text |Cite
|
Sign up to set email alerts
|

Self-supervised Dance Video Synthesis Conditioned on Music

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 48 publications
(29 citation statements)
references
References 22 publications
0
25
0
Order By: Relevance
“…Fig. 5 shows a group of translation results by using our method and previous state of the art methods on the music "Sorry" (also used in the previous work (Ren et al 2020)). It can be seen that the music-dance video generated by our method not only accurately capture the rhythm in the song, but also contain rich musical feelings and movement strength.…”
Section: Music-to-dance Translation Resultsmentioning
confidence: 99%
See 3 more Smart Citations
“…Fig. 5 shows a group of translation results by using our method and previous state of the art methods on the music "Sorry" (also used in the previous work (Ren et al 2020)). It can be seen that the music-dance video generated by our method not only accurately capture the rhythm in the song, but also contain rich musical feelings and movement strength.…”
Section: Music-to-dance Translation Resultsmentioning
confidence: 99%
“…Lee et al propose a decomposition-tocomposition framework for music-to-dance generation (Lee et al 2019), where they use a VAE to model dance units and use a Generative Adversarial Network (GAN) to organize the dance units based on input music. Ren et al integrate the local temporal discriminator and the global content discriminator for helping generate coherent dance sequences based on the noisy dataset, and then use pose-toappearance mapping to generate human dance videos (Ren et al 2020). However, all the above methods directly generate the dance movements from music, which inevitably leads to a problem of motion degradation and is not yet able to meet the requirements of expert-level music-to-dance translation.…”
Section: Music-to-dance Translationmentioning
confidence: 99%
See 2 more Smart Citations
“…For the reconstruction term in Eq. 3, we adopt a VGG perceptual loss (Simonyan & Zisserman, 2015;Ren et al, 2020), which is widely used in unsupervised disentanglement methods (Wu et al, 2020;2019c). For the Ψ-constraint, i.e.…”
Section: Proposed C-s Disentanglement Modulementioning
confidence: 99%