2018 13th IAPR International Workshop on Document Analysis Systems (DAS) 2018
DOI: 10.1109/das.2018.9
|View full text |Cite
|
Sign up to set email alerts
|

Encoding CNN Activations for Writer Recognition

Abstract: The encoding of local features is an essential part for writer identification and writer retrieval. While CNN activations have already been used as local features in related works, the encoding of these features has attracted little attention so far. In this work, we compare the established VLAD encoding with triangulation embedding. We further investigate generalized max pooling as an alternative to sum pooling and the impact of decorrelation and Exemplar SVMs. With these techniques, we set new standards on t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
46
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 44 publications
(46 citation statements)
references
References 31 publications
0
46
0
Order By: Relevance
“…Likewise, the ‘Numberberg’ system is based on exploiting CNN activations to characterise writer and is primarily based on the technique reported in Ref. [54]. Our proposed feature combination outperforms both these techniques indicating that ConvNets may not be able to learn robust feature representations due to the relatively limited amount of training data per class which is the case in most practical applications in the writer identification problem.…”
Section: Resultsmentioning
confidence: 99%
“…Likewise, the ‘Numberberg’ system is based on exploiting CNN activations to characterise writer and is primarily based on the technique reported in Ref. [54]. Our proposed feature combination outperforms both these techniques indicating that ConvNets may not be able to learn robust feature representations due to the relatively limited amount of training data per class which is the case in most practical applications in the writer identification problem.…”
Section: Resultsmentioning
confidence: 99%
“…They computed LeNet and ResNet based features and reported 99.5% on CVL, 99.6% on ICDAR13, KHATT in Top-1 respectively. In [18], LeNet and ResNet CNN models were used to learn the activation features. The activation features were utilized as local features by encoding with VLAD.…”
Section: Related Workmentioning
confidence: 99%
“…Manual features are language dependent. Automatic features learned by deep neural networks outperformed as compare to handcrafted features [18]- [21]. The Model based automatic features are extracted by deep learning based models automatically from the raw data of images directly.…”
Section: Introductionmentioning
confidence: 99%
“…In comparison, Christlein et al [3] proposed to use an unsupervised learning scheme to compute deep activation features that are eventually encoded using VLAD [26]. In a subsequent work [27], they show that GMP improves the encoding consistently. He et al [28] employ auxiliary tasks to improve writer identification of single word images.…”
Section: B Historical Document Image Classificationmentioning
confidence: 99%