2016 15th International Conference on Frontiers in Handwriting Recognition (ICFHR) 2016
DOI: 10.1109/icfhr.2016.0060
|View full text |Cite
|
Sign up to set email alerts
|

PHOCNet: A Deep Convolutional Neural Network for Word Spotting in Handwritten Documents

Abstract: In recent years, deep convolutional neural networks have achieved state of the art performance in various computer vision task such as classification, detection or segmentation. Due to their outstanding performance, CNNs are more and more used in the field of document image analysis as well. In this work, we present a CNN architecture that is trained with the recently proposed PHOC representation. We show empirically that our CNN architecture is able to outperform state of the art results for various word spot… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
206
0

Year Published

2018
2018
2019
2019

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 197 publications
(206 citation statements)
references
References 21 publications
0
206
0
Order By: Relevance
“…The core of the proposed method consists of using a CNN as a feature extractor. We have used PHOCnet [2], a CNN architecture recently proposed for segmentation-based KWS. PHOCnet was the best performing model on the recent ICFHR 2016 KWS competition (unpenalized MAP scenario) [1].…”
Section: Methods and Model Parametersmentioning
confidence: 99%
See 4 more Smart Citations
“…The core of the proposed method consists of using a CNN as a feature extractor. We have used PHOCnet [2], a CNN architecture recently proposed for segmentation-based KWS. PHOCnet was the best performing model on the recent ICFHR 2016 KWS competition (unpenalized MAP scenario) [1].…”
Section: Methods and Model Parametersmentioning
confidence: 99%
“…Regarding further details on the network architecture, as well as details on how training is performed (parameters, number of iterations, use of dropout, etc. ), the reader is referred to the original publication [2]. All layers between the input layer and the SPP layer are of variable size, as they depend on the input word image size.…”
Section: Neural Network Architecture and Deep Featuresmentioning
confidence: 99%
See 3 more Smart Citations