2019
DOI: 10.48550/arxiv.1903.10729
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

WGANSing: A Multi-Voice Singing Voice Synthesizer Based on the Wasserstein-GAN

Pritish Chandna,
Merlijn Blaauw,
Jordi Bonada
et al.
Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(9 citation statements)
references
References 0 publications
0
9
0
Order By: Relevance
“…Neural networks have been used previously to some success in modeling pre-extracted synthesis parameters (Blaauw & Bonada, 2017;Chandna et al, 2019), but these models fall short of endto-end learning. The analysis parameters must still be tuned by hand and gradients cannot flow through the synthesis procedure.…”
Section: Oscillator Modelsmentioning
confidence: 99%
“…Neural networks have been used previously to some success in modeling pre-extracted synthesis parameters (Blaauw & Bonada, 2017;Chandna et al, 2019), but these models fall short of endto-end learning. The analysis parameters must still be tuned by hand and gradients cannot flow through the synthesis procedure.…”
Section: Oscillator Modelsmentioning
confidence: 99%
“…Researches to extend the SVS system to the multi-singer system has been conducted relatively recently. [4] proposes a method of expressing each singer's identity by one-hot embedding. This method is straightforward and simple, but has the limitation of requiring re-training each time to add a new singer.…”
Section: Multi-singer Svs Systemmentioning
confidence: 99%
“…The multi-singer SVS system should not only produce natural pronunciation and pitch contour but also suitably reflect the identity of a particular singer. To achieve this, methods for adding conditional inputs reflecting the singer's identity to the network have been proposed [4,5].…”
Section: Introductionmentioning
confidence: 99%
“…Although there are few works focusing on the synthesis of Peking Opera, or more broadly, opera, the synthesis of singing voice has been researched since 1962 when Kelly and Lochbaum [1] used an acoustic tube model to synthesis singing voice with success. Recently, several works [2][3][4][5][6][7] use deep neural networks to synthesis singing voice which, known as parametric systems, process fundamental frequency (or pitch contour, f0) and harmonics features (or timbre) separately. As a typical case among such systems, Neural Parametric Singing Synthesizer (NPSS) [2] using a phoneme timing model, a pitch model and a timbre model each consist a set of neural networks * Yusong Wu performed the work while at Tencent.…”
Section: Introductionmentioning
confidence: 99%