2019 International Conference on Document Analysis and Recognition (ICDAR) 2019
DOI: 10.1109/icdar.2019.00220
|View full text |Cite
|
Sign up to set email alerts
|

Table Structure Extraction with Bi-Directional Gated Recurrent Unit Networks

Abstract: Tables present summarized and structured information to the reader, which makes table's structure extraction an important part of document understanding applications. However, table structure identification is a hard problem not only because of the large variation in the table layouts and styles, but also owing to the variations in the page layouts and the noise contamination levels. A lot of research has been done to identify table structure, most of which is based on applying heuristics with the aid of optic… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
36
0
2

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 54 publications
(38 citation statements)
references
References 20 publications
0
36
0
2
Order By: Relevance
“…We build the TFE as an attention bi-directional GRU network [38], [39] to recurrently process word embeddings in each element of the text list and adaptively merges them to element-level features based on attention scores. In particular, we first construct a bi-direction GRU network to process word embedding sequences from both directions.…”
Section: Duo Multi-modal Graph Convolutional Feature Encoder (Dgfe)mentioning
confidence: 99%
“…We build the TFE as an attention bi-directional GRU network [38], [39] to recurrently process word embeddings in each element of the text list and adaptively merges them to element-level features based on attention scores. In particular, we first construct a bi-direction GRU network to process word embedding sequences from both directions.…”
Section: Duo Multi-modal Graph Convolutional Feature Encoder (Dgfe)mentioning
confidence: 99%
“…Recurrent neural networks [40] have also been employed to handle the problem of table structure extraction [41], [42]. However, most of the prior approaches have utilized PDF meta-data.…”
Section: ) Recurrent Neural Networkmentioning
confidence: 99%
“…It can handle misspelled words and disambiguation problems also. LSTM [ 42 ], Bi-LSTM [ 43 ], GRU [ 44 ], and Bi-GRU [ 45 ] are investigated as classifiers (with one-dimensional convolution layer). These classifiers can process sequential data to overcome the short memory problem of recurrent neural networks.…”
Section: Corpus Collection and Dataset Statisticsmentioning
confidence: 99%
“…A Long-Short Term Memory (LSTM) worked in one sequence or forward direction. According to [4,45,62], Bi-LSTM can capture or calculate both directions of contexts, such as upcoming and previous hidden layers. Backward layer:…”
Section: Bidirectional Long-short Term Memory a Bidirectionalmentioning
confidence: 99%