2014
DOI: 10.1007/s11263-014-0793-6
|View full text |Cite
|
Sign up to set email alerts
|

Label Embedding: A Frugal Baseline for Text Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
53
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 105 publications
(54 citation statements)
references
References 43 publications
1
53
0
Order By: Relevance
“…Recently, there has been a lot of work surrounding the connection between image representations and word embeddings, both with natural images [8], and word images [1], [2], [9], [10]. For the handwritten word spotting community, hand-crafted image representations, primarily Fisher vectors and other bag-of-visual-words models, have been the features of choice [1], [3], [10].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Recently, there has been a lot of work surrounding the connection between image representations and word embeddings, both with natural images [8], and word images [1], [2], [9], [10]. For the handwritten word spotting community, hand-crafted image representations, primarily Fisher vectors and other bag-of-visual-words models, have been the features of choice [1], [3], [10].…”
Section: Related Workmentioning
confidence: 99%
“…Previous approaches use techniques like Canonical Correlation Analysis [1], Latent Semantic Analysis [9], or structured Support Vector Machines [9] to learn a mapping from image representation to a word embedding space. In order to allow for fine-tuning of the triplet-CNN, we opt for a small fully connected neural network, consisting of two hidden layers of size 4096.…”
Section: Learning the Embeddingmentioning
confidence: 99%
“…1 Traditionally, word spotting and recognition have focused on document images [6], [7], [8], [9], [10], [11], [12], [13], [14], where the main challenges come from differences in writing styles: the writing styles of different writers may be completely different for the same word. Recently, however, with the development of powerful computer vision techniques during the last decade, there has been an increased interest in performing word spotting and recognition on natural images [3], [4], [5], [15], [16], [17], which poses different challenges such as huge variations in illumination, point of view, typography, etc.…”
Section: Introductionmentioning
confidence: 99%
“…Furthermore, after CCA the attribute scores and binary attributes lie in a more correlated space, which makes the comparison between the scores and the PHOCs for our QBS problem more principled. CCA can also be seen as a label embedding method, similar in spirit to the recent approach of [27]. CCA is also used as a dimensionality reduction tool: we reduce the dimensionality from 384 down to 192-256 dimensions.…”
Section: Calibration Of Scoresmentioning
confidence: 99%
“…A naive implementation of this attributes representation greatly outperforms the direct use of FVs. A very similar string embedding has been simultaneously proposed in [27]. However, in their case, the representation is used in a label embedding context, and not as a source of attributes.…”
Section: Introductionmentioning
confidence: 99%