2014
DOI: 10.48550/arxiv.1412.2620
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Cells in Multidimensional Recurrent Neural Networks

Abstract: The transcription of handwritten text on images is one task in machine learning and one solution to solve it is using multi-dimensional recurrent neural networks (MDRNN) with connectionist temporal classification (CTC). The RNNs can contain special units, the long short-term memory (LSTM) cells. They are able to learn long term dependencies but they get unstable when the dimension is chosen greater than one. We defined some useful and necessary properties for the one-dimensional LSTM cell and extend them in th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
3
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 7 publications
0
3
0
Order By: Relevance
“…Additionally, we process multiple images and also the four directions of a multi-directional MDLSTM layer simultaneously by using batched cuBLAS operations and custom CUDA kernels. Optionally, the stable cell described in [31] can be used to improve convergence.…”
Section: Cuda Kernels For 1d and 2d Lstm Layersmentioning
confidence: 99%
“…Additionally, we process multiple images and also the four directions of a multi-directional MDLSTM layer simultaneously by using batched cuBLAS operations and custom CUDA kernels. Optionally, the stable cell described in [31] can be used to improve convergence.…”
Section: Cuda Kernels For 1d and 2d Lstm Layersmentioning
confidence: 99%
“…Thus, we use the RNN with LSTM mechanism extended to the 2-D case named 2-D LSTM [28], [29] Similar with some machine translation tasks [30], [31], the compression and reconstruction of CSI is essentially a feature extraction and representation task from one 2-D sequence to one 2-D sequence. Therefore, we propose a 2-D Sequence to Sequence (Seq2Seq) neural network mainly composed of 2-D LSTMs for the CSI feedback task.…”
mentioning
confidence: 99%
“…2-D RNN and 2-D LSTM Mechanism for MIMO-OFDM CSITo deal with sequence-dependent data, RNN shows superior advantage due to its recurrent calculation structure. For data with multi-dimensional (M-D) sequence-dependence, the authors in[28],[29] introduced M-D RNN and M-D LSTM as a generalization of standard RNN and standard LSTM, respectively. Through using a M-D recurrent calculation structure, M-D RNN can overcome the limitation of one-dimensional (1-D) RNN, which can only utilize 1-D sequence dependence.…”
mentioning
confidence: 99%