2015
DOI: 10.1007/s00034-015-0134-1
|View full text |Cite
|
Sign up to set email alerts
|

Vowel-Based Non-uniform Prosody Modification for Emotion Conversion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 20 publications
(5 citation statements)
references
References 13 publications
0
5
0
Order By: Relevance
“…These features were used to build emotion detection [206] and emotion recognition systems [19], [20], [205]. The effectiveness of excitation during the production of emotional speech was examined in [207]- [209] using prosody modification to convert neutral speech to emotional speech.…”
Section: B Study Of Vocal Emotionsmentioning
confidence: 99%
“…These features were used to build emotion detection [206] and emotion recognition systems [19], [20], [205]. The effectiveness of excitation during the production of emotional speech was examined in [207]- [209] using prosody modification to convert neutral speech to emotional speech.…”
Section: B Study Of Vocal Emotionsmentioning
confidence: 99%
“…Prosody features, such as pitch contour, jitter [21], [30], and speaking rate [47] have been analyzed in literature. Prosody modification methods mainly focus on either static (sentencewise) or dynamic (region-wise) prosody transformation [22], [24], [47], [48]. In most of these, epoch locations, which are the instants of glottal closure, have been taken as anchor points for prosody manipulation.…”
Section: Related Workmentioning
confidence: 99%
“…MCD is used to evaluate the spectral conversion efficiency in emotion conversion [45], [55], [56]. It is calculated as in (24)…”
Section: ) MCDmentioning
confidence: 99%
See 1 more Smart Citation
“…Recently, the importance of features extracted around GCIs has been observed in several studies in emotional speech analysis and detection [17][18][19]. The importance of frequency of vibration of vocal folds (interval between GCIs) was studied in emotion detection and emotion conversion/synthesis in [18,20]. In addition, some speech synthesis systems (e.g., [21][22][23]) use voice source modeling and GCI detection in generation of the synthesizer's excitation signal.…”
Section: Introductionmentioning
confidence: 99%