2021
DOI: 10.1007/978-3-030-86331-9_37
|View full text |Cite
|
Sign up to set email alerts
|

Handwritten Mathematical Expression Recognition with Bidirectionally Trained Transformer

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

1
32
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 53 publications
(40 citation statements)
references
References 27 publications
1
32
0
Order By: Relevance
“…In order to be consistent with the reported inference process, DWAP-TD [42] and BTTR [46] use beam search, while DWAP [39] doesn't. Specifically, as shown in table 3, our method outperforms BTTR [46] by 1.6% on easy subset. However, as the difficulty of the test subset increases, the leading margin of our method increases to 5.5% on the hard subset.…”
Section: Comparisons With State-of-the-artsmentioning
confidence: 99%
See 3 more Smart Citations
“…In order to be consistent with the reported inference process, DWAP-TD [42] and BTTR [46] use beam search, while DWAP [39] doesn't. Specifically, as shown in table 3, our method outperforms BTTR [46] by 1.6% on easy subset. However, as the difficulty of the test subset increases, the leading margin of our method increases to 5.5% on the hard subset.…”
Section: Comparisons With State-of-the-artsmentioning
confidence: 99%
“…Most HMER methods extensively adopt the sequence-to-sequence approach. The authors in [8,15,16,23,27,29,39,40,40,43,46] proposed an attention-based sequence-to-sequence model to convert the handwritten mathematical expression images into rep- resentational markup language LaTeX. Recently, Wu et al [31] designed a graph-to-graph(G2G) model that explores the HMEs structural relationship of the input formula and output markup, which significantly improve the performance.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…The transformer architecture [5] was first introduced in the context of machine translation. Since then, transformerbased models have proven their robustness through a wide variety of computer vision tasks such as images of mathematical expression recognition [6] or image classification [7]. They have also shown promising results for single text line recognition [8], [9] and scene text recognition [10].…”
mentioning
confidence: 99%