2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01004
|View full text |Cite
|
Sign up to set email alerts
|

Sign Language Transformers: Joint End-to-End Sign Language Recognition and Translation

Abstract: Prior work on Sign Language Translation has shown that having a mid-level sign gloss representation (effectively recognizing the individual signs) improves the translation performance drastically. In fact, the current state-of-theart in translation requires gloss level tokenization in order to work. We introduce a novel transformer based architecture that jointly learns Continuous Sign Language Recognition and Translation while being trainable in an end-to-end manner. This is achieved by using a Connectionist … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
119
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 264 publications
(165 citation statements)
references
References 44 publications
2
119
0
Order By: Relevance
“…However, this claim assumes that the ground truth gloss annotations give a full understanding of sign language, which ignores the information bottleneck in glosses. Camgoz et al (2020) hypothesizes that it is therefore possible to surpass G2T performance without using GT glosses, which we confirm in this section.…”
Section: German Sign2gloss2text (S2g2t)supporting
confidence: 83%
See 4 more Smart Citations
“…However, this claim assumes that the ground truth gloss annotations give a full understanding of sign language, which ignores the information bottleneck in glosses. Camgoz et al (2020) hypothesizes that it is therefore possible to surpass G2T performance without using GT glosses, which we confirm in this section.…”
Section: German Sign2gloss2text (S2g2t)supporting
confidence: 83%
“…Also, we report an improvement of over 5 BLEU-4 on the state-of-the-art. A single Transformer also gives an improvement of over 4 BLEU-4 more than the state-of-the-art, which shows the advantage of Transformers for SLT, as shown also in Camgoz et al (2020). We also use 5 of the best models from our experiments on ASLG-PC12 in an ensemble.…”
Section: Ensemble Decodingmentioning
confidence: 71%
See 3 more Smart Citations