International Conference on Acoustics, Speech, and Signal Processing
DOI: 10.1109/icassp.1990.115768
|View full text |Cite
|
Sign up to set email alerts
|

A hybrid coder for hidden Markov models using a recurrent neural networks

Abstract: A hybrid coder is introduced for obtaining descriptions of speech patterns, This coder uses popular Vector Quantization (VQ) techniques on melscale cepstral coefficients and their derivatives together with a Recurrent Network (RN) for describing suprasegmental features of speech.The purpose of these features is to focus the search when Hidden Markov Models (HMM) are used for speech unit or word models.Preliminary experiments of speakerindependent connected digit recognition showed that using a hybrid coder bas… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0
1

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 8 publications
0
5
0
1
Order By: Relevance
“…For the DNN architecture, we used 4 hidden layers, each with 2048 units and a Rectified Linear Unit (ReLU) [92] activation function. For classification, we used a softmax layer [93], which assigns a probability to each class, i.e., a probability to each HMM state. The system's architecture is depicted in Figure 6.…”
Section: Deep Neural Network For Acoustic Modellingmentioning
confidence: 99%
“…For the DNN architecture, we used 4 hidden layers, each with 2048 units and a Rectified Linear Unit (ReLU) [92] activation function. For classification, we used a softmax layer [93], which assigns a probability to each class, i.e., a probability to each HMM state. The system's architecture is depicted in Figure 6.…”
Section: Deep Neural Network For Acoustic Modellingmentioning
confidence: 99%
“…A variety of solutions have been proposed to tackle this problem. One solution uses the ANN to compute an additional set of symbols as transformed observations for the HMM [90]. A further improvement of this method is achieved through a global optimization of both the ANN and HMM [91].…”
Section: Proposed Solutionsmentioning
confidence: 99%
“…Otra posible solución es utilizar redes neuronales para computar conjuntos de símbolos adicionales, que puedan entregarse como observaciones transformadas para los HMM [BEN90]. Más específicamente, la red genera grados de certeza para los rasgos básicos de los sonidos como sonoridad, fricación y oclusión/silencio.…”
Section: Rasgounclassified