2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.01173
|View full text |Cite
|
Sign up to set email alerts
|

Stochastic Transformer Networks with Linear Competing Units: Application to end-to-end SL Translation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 20 publications
(12 citation statements)
references
References 24 publications
0
7
0
Order By: Relevance
“…Camgöz et al mention that the most common errors are related to numbers, dates, and places: these can be difficult to derive from context in weather broadcasts [6,37]. The same kind of errors is made by the models of Partaourides et al [70] and Voskou et al [81]. Zheng et al illustrate how their model improves accuracy for longer sentences [64].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Camgöz et al mention that the most common errors are related to numbers, dates, and places: these can be difficult to derive from context in weather broadcasts [6,37]. The same kind of errors is made by the models of Partaourides et al [70] and Voskou et al [81]. Zheng et al illustrate how their model improves accuracy for longer sentences [64].…”
Section: Discussionmentioning
confidence: 99%
“…Six papers use features extracted using a 2D CNNs by first training a CSLR model on RWTH-PHOENIX-Weather 2014 6 [6,36,38,81,92,93]. These papers use the full frame as inputs to the feature extractor.…”
Section: Sign Language Representationsmentioning
confidence: 99%
“…In the related LWTA literature, the competition process is usually deterministic, i.e., the unit with the highest linear activation is deemed the winner each time. However, novel data-driven stochastic arguments for the competition process have been recently proposed in Theodoridis 2019, 2021;Panousis et al 2021;Voskou et al 2021).…”
Section: Foundational Principlesmentioning
confidence: 99%
“…Training was performed collaboratively for the entire system (both tasks). The need for that intermediate step has been alleviated in later works such as [ 10 ], where a winner-takes-all activation is integrated into the Transformer architecture. In [ 35 ], the authors introduced a context-aware continuous sign language recognition using a generative adversarial network architecture.…”
Section: Related Workmentioning
confidence: 99%
“…Recent methods based on networks with self-attention (transformers) [ 9 , 10 ], which currently represent the state-of-the-art in SLT, have yielded promising results owing to their ability to learn without depending too much on expert knowledge. Nevertheless, to fully unleash their performance and generalization potential, such systems require large corpora for training, which increase with the size of the vocabulary.…”
Section: Introductionmentioning
confidence: 99%