5th International Conference on Spoken Language Processing (ICSLP 1998) 1998
DOI: 10.21437/icslp.1998-147
|View full text |Cite
|
Sign up to set email alerts
|

Emotional speech synthesis: from speech database to TTS

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
3
0
2

Year Published

2001
2001
2023
2023

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 55 publications
(6 citation statements)
references
References 6 publications
0
3
0
2
Order By: Relevance
“…Over the last several decades, there has been an emphasis on emotion representation in speech. Montero et al [ 22 ] put forward that synthesised speech cannot be marked as natural sounding in the absence of emotional features. Li and Zhao [ 23 ] used acoustic features to identify emotions in speech.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Over the last several decades, there has been an emphasis on emotion representation in speech. Montero et al [ 22 ] put forward that synthesised speech cannot be marked as natural sounding in the absence of emotional features. Li and Zhao [ 23 ] used acoustic features to identify emotions in speech.…”
Section: Literature Reviewmentioning
confidence: 99%
“…El producto de esta interacción hablante humano-ChatGPT, la «lengua-e» (Mendívil 2023) en tanto que externa, puede analizarse con las herramientas que nos proporcionan disciplinas lingüísticas como el análisis del discurso, la lingüística textual o el análisis de la conversación. Contamos ya con veinticinco años de investigaciones que analizan, desde el campo de la lingüística, las interacciones humano-máquina, como la de Lee y Narayanan (2005), que analizan las conversaciones con un call center, las de Montero et al (1998a;1998b) para el discurso emotivo, o Přibil y Přibilová (2010), que recurren a actores para grabar párrafos, frases cortas y palabras aisladas con el objetivo de aplicar los resultados al discurso robotizado (Text-To-Speech System). Igualmente, las investigaciones de Fischer (2023) y Fischer y Matsumoto (2023), sobre las interacciones con robots sociales o sobre el discurso persuasivo de estos (Fischer, Fischer y Palinko 2023;Fucinato, Niebuhr, Nørskov y Fischer 2023;Langedijk y Fischer 2023).…”
Section: La Inteligencia Artificial Y Los Modelos De Lenguajeunclassified
“…The average recognition rates achieved were 92.5% and 90% for DBN and BP, respectively. In addition, in [32], restricted Boltzmann machines (RBMs) and DBN were used together with audio files from one female Spanish speaker from the emotional speech dataset [33] as part of large project, INTERFACE, with the big six classes, joy, sadness, anger, fear, disgust and surprise along with neutral. Those authors used two kinds of features for extraction, MFCC and prosodic features with RBM and DBN, providing a maximum classification error rate of 40.82%…”
Section: Classification Based On Six Classesmentioning
confidence: 99%