2011
DOI: 10.1016/j.specom.2011.05.001
|View full text |Cite
|
Sign up to set email alerts
|

Speaker-independent HMM-based voice conversion using adaptive quantization of the fundamental frequency

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2012
2012
2021
2021

Publication Types

Select...
4
4

Relationship

2
6

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 24 publications
0
5
0
Order By: Relevance
“…One interesting direction to investigate, in the context of human-animal interaction and sound design, is how to generate call sequences from human vocal imitation or sketching (e.g. analogously to [14], build a parallel database of animal vocalization and human imitations, and train models for both), as well as methods to embed human emotions in the synthetic vocalizations.…”
Section: Discussionmentioning
confidence: 99%
“…One interesting direction to investigate, in the context of human-animal interaction and sound design, is how to generate call sequences from human vocal imitation or sketching (e.g. analogously to [14], build a parallel database of animal vocalization and human imitations, and train models for both), as well as methods to embed human emotions in the synthetic vocalizations.…”
Section: Discussionmentioning
confidence: 99%
“…The reference sample was vocoded speech of the target speaker. Other detailed experimental conditions are found in [11]. The results are shown in Fig.…”
Section: Very Low Bit-rate Speech Coding Based On Msd-hmm With Qumentioning
confidence: 99%
“…A speaker-independent hidden Markov model (HMM) -based voice conversion technique was proposed by Nose and Kobayashi. The study included context-dependent prosodic symbols obtained using adaptive quantization of the fundamental frequency (F0) [63]. The input utterance of a source speaker was decoded into phonetic and prosodic symbol sequences and the converted speech was generated using the decoded information.…”
Section: Published Work In the Year 2011mentioning
confidence: 99%