2012
DOI: 10.1007/978-3-642-24797-2
|View full text |Cite
|
Sign up to set email alerts
|

Supervised Sequence Labelling with Recurrent Neural Networks

Abstract: Recurrent neural networks are powerful sequence learners. They are able to incorporate context information in a flexible way, and are robust to localised distortions of the input data. These properties make them well suited to sequence labelling, where input sequences are transcribed with streams of labels. Long short-term memory is an especially promising recurrent architecture, able to bridge long time delays between relevant input and output events, and thereby access long range context. The aim of this the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

7
2,088
0
15

Year Published

2012
2012
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 2,616 publications
(2,262 citation statements)
references
References 121 publications
7
2,088
0
15
Order By: Relevance
“…Parameters of LSTM are trained using BPTT. Core structure of LSTM is illustrated as follows (Graves, 2012): Figure 3. Structure of LSTM memory block…”
Section: Rnn and Lstmmentioning
confidence: 99%
“…Parameters of LSTM are trained using BPTT. Core structure of LSTM is illustrated as follows (Graves, 2012): Figure 3. Structure of LSTM memory block…”
Section: Rnn and Lstmmentioning
confidence: 99%
“…RNN has been widely used for sequence generation tasks (Graves, 2012a;Schuster and Paliwal, 1997). RNN accepts sequence of inputs X = {x 1 , x 2 , x 3 , ..., x |X| }, and gets h t at time t according to Equation (2).…”
Section: Rnnmentioning
confidence: 99%
“…Since these gates allow for write, read, and reset operations within a memory block, an LSTM block can be interpreted as (differentiable) memory chip in a digital computer. The overall effect of the gate units is that the LSTM memory cells can store and access information over long periods of time and thus avoid the vanishing gradient problem (for details see [6]). …”
Section: Bidirectional Long Short-term Memorymentioning
confidence: 99%
“…To extend the Jacobian to recurrent neural networks, we have to specify the timesteps (representing utterances) at which the input and output variables are measured. Thus, we calculate a four-dimensional matrix called the sequential Jacobian [6] to determine the sensitivity of the network outputs at time t to the inputs at time t :…”
Section: Sequential Jacobian Analysismentioning
confidence: 99%
See 1 more Smart Citation