2020
DOI: 10.1186/s13640-020-00523-5
|View full text |Cite
|
Sign up to set email alerts
|

Detection and recognition of cursive text from video frames

Abstract: Textual content appearing in videos represents an interesting index for semantic retrieval of videos (from archives), generation of alerts (live streams), as well as high level applications like opinion mining and content summarization. The key components of such systems require detection and recognition of textual content which also make the subject of our study. This paper presents a comprehensive framework for detection and recognition of textual content in video frames. More specifically, we target cursive… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(6 citation statements)
references
References 115 publications
0
5
0
Order By: Relevance
“…Experiments are conducted on 12,000 text lines culled from 4,000 video frames from Pakistani news stations. Reference [7] proposed a comparable UrduNet model, which is a hybrid approach of CNN and LSTM. On a self-generated dataset of almost 13,000 frames, a complete set of experiments are carried out.…”
Section: Background and Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Experiments are conducted on 12,000 text lines culled from 4,000 video frames from Pakistani news stations. Reference [7] proposed a comparable UrduNet model, which is a hybrid approach of CNN and LSTM. On a self-generated dataset of almost 13,000 frames, a complete set of experiments are carried out.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Mirza et al [7] present an implicit technique for new tickers from video frames by combining CNN and LSTM. This can be seen that even on its own training dataset, the model has difficulty in training with low accuracy.…”
Section: Explicit Segmentationmentioning
confidence: 99%
“…Naz et al [27] extracts features with CNN by the usage of MNIST dataset and does not exactly contemplate Urdu text features. In addition, as mentioned prior [14] Experiments and results along with comparisons with existing systems for text recognition are presented in Section IV, whereas section V concludes the study.…”
Section: ) Explicit Segmentationmentioning
confidence: 99%
“…Tesseract is an open-source OCR engine it takes images that attempt to acknowledge the text. The result is a text string, and the degree to which an image is similar to a human-readable text is measured by its correctness [17]. The OCR is used to recognize printed documents in papers, handwritten characters, or physical text messages such as license plates, street signs, and street numbers, document account holders, legal forms, ID cards, and so on.…”
Section: Feature Extractionmentioning
confidence: 99%