2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR) 2017
DOI: 10.1109/icdar.2017.361
|View full text |Cite
|
Sign up to set email alerts
|

Implicit Language Model in LSTM for OCR

Abstract: Neural networks have become the technique of choice for OCR, but many aspects of how and why they deliver superior performance are still unknown. One key difference between current neural network techniques using LSTMs and the previous state-of-the-art HMM systems is that HMM systems have a strong independence assumption. In comparison LSTMs have no explicit constraints on the amount of context that can be considered during decoding. In this paper we show that they learn an implicit LM and attempt to character… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 34 publications
(24 citation statements)
references
References 19 publications
(21 reference statements)
0
24
0
Order By: Relevance
“…Transkribus [34,35] is another complex platform for analysis of historical documents which covers many research areas such as layout analysis and handwritten text recognition. It includes also OCR using ABBYY Finereader Engine 11 6 . To the best of our knowledge, Tesseract and Transcribus are the best performing OCR systems.…”
Section: Existing Tools and Ocr Systemsmentioning
confidence: 99%
See 2 more Smart Citations
“…Transkribus [34,35] is another complex platform for analysis of historical documents which covers many research areas such as layout analysis and handwritten text recognition. It includes also OCR using ABBYY Finereader Engine 11 6 . To the best of our knowledge, Tesseract and Transcribus are the best performing OCR systems.…”
Section: Existing Tools and Ocr Systemsmentioning
confidence: 99%
“…Language models are often used in the OCR field to correct recognition errors [41]. Sabir et al [6] showed that LSTMbased models are able to learn language model implicitly. This LM is trained during the learning of the whole network.…”
Section: Impact Of the Implicit Language Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…They enable to share the contextual information between the locations and therefore enhance the performances. Nevertheless, it is known [22] that they can also learn some kind of language modeling.…”
Section: Impact Of Language Modelingmentioning
confidence: 99%
“…This prevents explicit learning of query selection. However, it is known that recurrent networks can learn implicit tasks [28]. Second, query selection is a soft decision made by the model, whereas a discrete decision is preferred.…”
Section: Appendix a Recurrent Neural Network Architecturementioning
confidence: 99%