2005
DOI: 10.1016/j.neunet.2005.06.042
|View full text |Cite
|
Sign up to set email alerts
|

Framewise phoneme classification with bidirectional LSTM and other neural network architectures

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

11
2,065
0
20

Year Published

2007
2007
2022
2022

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 4,328 publications
(2,279 citation statements)
references
References 5 publications
11
2,065
0
20
Order By: Relevance
“…Furthermore, the complexity of a task, and therefore the number of weights likely to be needed for it, does not necessarily increase with the dimensionality of the data. For example, both the networks described in this paper have less than half the weights than the one dimensional networks we have previously applied to speech recognition [4,3]. For a given task, we have also found that using a multi-directional MDRNN gives better results than a uni-directional MDRNN with the same overall number of weights, as previously demonstrated in one dimension [4].…”
Section: Multi-directional Mdrnnssupporting
confidence: 73%
“…Furthermore, the complexity of a task, and therefore the number of weights likely to be needed for it, does not necessarily increase with the dimensionality of the data. For example, both the networks described in this paper have less than half the weights than the one dimensional networks we have previously applied to speech recognition [4,3]. For a given task, we have also found that using a multi-directional MDRNN gives better results than a uni-directional MDRNN with the same overall number of weights, as previously demonstrated in one dimension [4].…”
Section: Multi-directional Mdrnnssupporting
confidence: 73%
“…More precisely, three connected handwriting competitions at ICDAR 2009 in three different languages (French, Arab, Farsi) were won by deep LSTM RNNs without any a priori linguistic knowledge, performing simultaneous segmentation and recognition. Compare (Graves and Schmidhuber, 2005;Schmidhuber et al, 2011;Graves et al, 2013;Graves and Jaitly, 2014) (Sec. 5.22).…”
Section: : First Official Competitions Won By Rnns and With Mpcnnsmentioning
confidence: 99%
“…As the reversed order also takes useful information, a backward representation can be achieved by feeding LSTM with the same input in reverse. We adopt the concatenation of the forward and backward LSTMs outputs, referred to as bidirectional LSTM (Graves and Schmidhuber, 2005). Figure 3a shows the neural network architecture of our E-E, E-T classifier.…”
Section: Temporal Relation Classifiersmentioning
confidence: 99%