2020
DOI: 10.1007/978-3-030-58517-4_11
|View full text |Cite
|
Sign up to set email alerts
|

Stochastic Fine-Grained Labeling of Multi-state Sign Glosses for Continuous Sign Language Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
41
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 60 publications
(45 citation statements)
references
References 17 publications
0
41
0
Order By: Relevance
“…Instead, it provided extra supervision and assisted the CSLR network to learn better gloss alignments. Niu et al in [ 43 ], proposed a 2D-CNN followed by a Transformer network for CSLR. They used three stochastic methods to drop frames of the input video, to randomly stop gradients of back-propagation and to model glosses using hidden states, respectively, which led to better CSLR performance.…”
Section: Sign Language Recognitionmentioning
confidence: 99%
“…Instead, it provided extra supervision and assisted the CSLR network to learn better gloss alignments. Niu et al in [ 43 ], proposed a 2D-CNN followed by a Transformer network for CSLR. They used three stochastic methods to drop frames of the input video, to randomly stop gradients of back-propagation and to model glosses using hidden states, respectively, which led to better CSLR performance.…”
Section: Sign Language Recognitionmentioning
confidence: 99%
“…Likewise, another continuous-SLR system using deep-CNNs and bidirectional recurrent neural networks for sequence learning can be found in [5]. A stochastic modeling of sign language components is proposed in [26] using a transformer encoder and connectionist temporal classification decoder. A modified-LSTM framework is discussed in [23] where the authors used a probability threshold to reset the LSTM's cell.…”
Section: Related Workmentioning
confidence: 99%
“…It inspires more research groups to study continuous SLR with large-scale vocabulary [13,22,34]. To enable end-to-end training, connectionist temporal classification (CTC) [15] is widely adopted by continuous SLR models [7,14,30,33,49]. With the development of neural machine translation, Camgöz et al formulate a new task, neural sign language translation (SLT) [10], which is becoming an active and promising direction [11,25].…”
Section: Related Workmentioning
confidence: 99%
“…The sign embedding layer is trained endto-end under CTC Loss with batch size 2. No iterative training [53], online refining [7] or temporal sampling [30] are used. We use Adam optimizer [21] and set the weight decay to 1 × 10 −6 .…”
Section: Implementation Detailsmentioning
confidence: 99%