2022
DOI: 10.3390/info13050220
|View full text |Cite
|
Sign up to set email alerts
|

Leveraging Frozen Pretrained Written Language Models for Neural Sign Language Translation

Abstract: We consider neural sign language translation: machine translation from signed to written languages using encoder–decoder neural networks. Translating sign language videos to written language text is especially complex because of the difference in modality between source and target language and, consequently, the required video processing. At the same time, sign languages are low-resource languages, their datasets dwarfed by those available for written languages. Recent advances in written language processing a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 18 publications
(23 citation statements)
references
References 31 publications
(55 reference statements)
0
22
0
1
Order By: Relevance
“…20 papers use a 2D CNN as feature extractor [6, 35-38, 64, 65, 72, 75, 77-81, 87, 88, 92, 93, 95, 98]. These are often pre-trained for image classification using the ImageNet dataset [274]; some are further pre-trained on the task of Continuous Sign Language Recognition (CSLR), e.g., [37,38,92,93]. Three papers use a subsequent 1D CNN to temporally process the resulting spatial features [64,77,80].…”
Section: Extraction Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…20 papers use a 2D CNN as feature extractor [6, 35-38, 64, 65, 72, 75, 77-81, 87, 88, 92, 93, 95, 98]. These are often pre-trained for image classification using the ImageNet dataset [274]; some are further pre-trained on the task of Continuous Sign Language Recognition (CSLR), e.g., [37,38,92,93]. Three papers use a subsequent 1D CNN to temporally process the resulting spatial features [64,77,80].…”
Section: Extraction Methodsmentioning
confidence: 99%
“…For instance, Koller et al presented SLR systems that exploit SignWriting [33,34], and these systems are leveraged in some later works on SLT, e.g., [35,36]. Many SLT models also use feature extractors that were pre-trained with gloss labels, e.g., [37,38].…”
Section: Notation Systems For Sign Languagesmentioning
confidence: 99%
See 3 more Smart Citations