Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.01090
|View full text |Cite
|
Sign up to set email alerts
|

SignBERT: Pre-Training of Hand-Model-Aware Representation for Sign Language Recognition

Abstract: Hand gesture serves as a crucial role during the expression of sign language. Current deep learning based methods for sign language understanding (SLU) are prone to over-fitting due to insufficient sign data resource and suffer limited interpretability. In this paper, we propose the first self-supervised pre-trainable SignBERT+ framework with model-aware hand prior incorporated. In our framework, the hand pose is regarded as a visual token, which is derived from an off-the-shelf detector. Each visual token is … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 44 publications
(35 citation statements)
references
References 88 publications
(157 reference statements)
0
10
0
Order By: Relevance
“…As shown in Table 5, we compare with the previous methods. SignBERT (Hu et al 2021a) Our method outperforms SignBERT (Hu et al 2021a) with 4.89%, 5.96% and 9.28% Top-1 per-instance accuracy improvement on MSASL100, MSASL200 and MSASL1000, respectively. Notably, our method even achieves comparable performance with RGB-based methods.…”
Section: Comparison With State-of-the-art Methodsmentioning
confidence: 79%
See 2 more Smart Citations
“…As shown in Table 5, we compare with the previous methods. SignBERT (Hu et al 2021a) Our method outperforms SignBERT (Hu et al 2021a) with 4.89%, 5.96% and 9.28% Top-1 per-instance accuracy improvement on MSASL100, MSASL200 and MSASL1000, respectively. Notably, our method even achieves comparable performance with RGB-based methods.…”
Section: Comparison With State-of-the-art Methodsmentioning
confidence: 79%
“…Method MSASL100 MSASL200 MSASL1000 P-I P-C P-I P-C P-I (Yan, Xiong, and Lin 2018) 59 (Yan, Xiong, and Lin 2018) 90.0 SignBERT (Hu et al 2021a) 94.5 Ours 95.4…”
Section: Ablation Studymentioning
confidence: 99%
See 1 more Smart Citation
“…Sign language recognition methods can be roughly categorized into isolated sign language recognition [19,20,44] and continuous sign language recognition [5,7,34,35,38] (CSLR), and we focus on the latter in this paper. CSLR tries to translate image frames into corresponding glosses in a weakly-supervised way: only sentence-level label is provided.…”
Section: Continuous Sign Language Recognitionmentioning
confidence: 99%
“…In the domain of continuous sign language recognition, in which the objective is to infer a sequence of sign glosses, prior work has explored HMMs [3,36] in combination with Dynamic Time Warping (DTW) [73], RNNs [18] and architectures capable of learning effectively from CTC losses [15,75]. Recently, sign representation learning methods inspired by BERT [20] have shown the potential to learn effective representations for both isolated [24] and continuous [76] recognition. Koller [35] provides an extensive survey of the sign recognition literature, highlighting the extremely limited supply of datasets with large-scale vocabularies suitable for continuous sign language recognition.…”
Section: Related Workmentioning
confidence: 99%