1994
DOI: 10.1109/72.279192
|View full text |Cite
|
Sign up to set email alerts
|

An application of recurrent nets to phone probability estimation

Abstract: This paper presents an application of recurrent networks for phone probability estimation in large vocabulary speech recognition. The need for efficient exploitation of context information is discussed; a role for which the recurrent net appears suitable. An overview of early developments of recurrent nets for phone recognition is given along with the more recent improvements that include their integration with Markov models. Recognition results are presented for the DARPA TIMIT and Resource Management tasks, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
160
0
4

Year Published

1999
1999
2012
2012

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 367 publications
(168 citation statements)
references
References 27 publications
4
160
0
4
Order By: Relevance
“…Hybrids of hidden Markov models (HMMs) and artificial neural networks (ANNs) were proposed by several researchers in the 1990s as a way of overcoming the drawbacks of HMMs (Bourlard and Morgan, 1994;Bengio, 1993;Robinson, 1994;Bengio, 1999). The introduction of ANNs was intended to provide more discriminant training, improved modelling of phoneme duration, richer, nonlinear function approximation, and perhaps most importantly, increased use of contextual information.…”
Section: Introductionmentioning
confidence: 99%
“…Hybrids of hidden Markov models (HMMs) and artificial neural networks (ANNs) were proposed by several researchers in the 1990s as a way of overcoming the drawbacks of HMMs (Bourlard and Morgan, 1994;Bengio, 1993;Robinson, 1994;Bengio, 1999). The introduction of ANNs was intended to provide more discriminant training, improved modelling of phoneme duration, richer, nonlinear function approximation, and perhaps most importantly, increased use of contextual information.…”
Section: Introductionmentioning
confidence: 99%
“…This system was tested with the Wall Street Journal database. The TIMIT results came from a hybrid RNN/HMM in 1994, (Robinson, 1994). The inputs to the neural network are features extracted using a long left context.…”
Section: Overview Of Current and Past Research On Timit Phone Recognimentioning
confidence: 99%
“…Spoken utterances are represented as arrays of phoneme probabilities. A recurrent neural network similar to [24] processes RASTA-PLP coefficients [25] to estimate phoneme and speech/silence probabilities. The RNN has 12 input units, 176 hidden units, and 40 output units.…”
Section: Representing and Comparing Spoken Utterancesmentioning
confidence: 99%